Store log summary data in Redis instead of Firebase
This commit is contained in:
parent
6659746b3d
commit
05fde65cd3
|
@ -16,3 +16,5 @@ npm-debug.log*
|
||||||
yarn-debug.log*
|
yarn-debug.log*
|
||||||
yarn-error.log*
|
yarn-error.log*
|
||||||
|
|
||||||
|
# redis
|
||||||
|
dump.rdb
|
||||||
|
|
|
@ -1,15 +1,13 @@
|
||||||
require('isomorphic-fetch')
|
require('isomorphic-fetch')
|
||||||
const parseURL = require('url').parse
|
const parseURL = require('url').parse
|
||||||
const crypto = require('crypto')
|
|
||||||
const invariant = require('invariant')
|
const invariant = require('invariant')
|
||||||
const admin = require('firebase-admin')
|
|
||||||
const gunzip = require('gunzip-maybe')
|
const gunzip = require('gunzip-maybe')
|
||||||
const ndjson = require('ndjson')
|
const ndjson = require('ndjson')
|
||||||
|
const redis = require('redis')
|
||||||
|
|
||||||
const CloudflareEmail = process.env.CLOUDFLARE_EMAIL
|
const CloudflareEmail = process.env.CLOUDFLARE_EMAIL
|
||||||
const CloudflareKey = process.env.CLOUDFLARE_KEY
|
const CloudflareKey = process.env.CLOUDFLARE_KEY
|
||||||
const FirebaseURL = process.env.FIREBASE_URL
|
const RedisURL = process.env.REDIS_URL
|
||||||
const FirebaseAccount = process.env.FIREBASE_ACCOUNT
|
|
||||||
|
|
||||||
invariant(
|
invariant(
|
||||||
CloudflareEmail,
|
CloudflareEmail,
|
||||||
|
@ -22,13 +20,8 @@ invariant(
|
||||||
)
|
)
|
||||||
|
|
||||||
invariant(
|
invariant(
|
||||||
FirebaseURL,
|
RedisURL,
|
||||||
'Missing the $FIREBASE_URL environment variable'
|
'Missing the $REDIS_URL environment variable'
|
||||||
)
|
|
||||||
|
|
||||||
invariant(
|
|
||||||
FirebaseAccount,
|
|
||||||
'Missing the $FIREBASE_ACCOUNT environment variable'
|
|
||||||
)
|
)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -55,17 +48,7 @@ Stuff we wanna show on the website:
|
||||||
- Browser usage
|
- Browser usage
|
||||||
*/
|
*/
|
||||||
|
|
||||||
const serviceAccount = JSON.parse(FirebaseAccount)
|
const db = redis.createClient(RedisURL)
|
||||||
|
|
||||||
admin.initializeApp({
|
|
||||||
credential: admin.credential.cert(serviceAccount),
|
|
||||||
databaseURL: FirebaseURL,
|
|
||||||
databaseAuthVariableOverride: {
|
|
||||||
uid: 'ingest-logs-worker'
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
const db = admin.database()
|
|
||||||
|
|
||||||
const getZones = (domain) =>
|
const getZones = (domain) =>
|
||||||
fetch(`https://api.cloudflare.com/client/v4/zones?name=${domain}`, {
|
fetch(`https://api.cloudflare.com/client/v4/zones?name=${domain}`, {
|
||||||
|
@ -91,10 +74,7 @@ const toSeconds = (millis) =>
|
||||||
Math.floor(millis / 1000)
|
Math.floor(millis / 1000)
|
||||||
|
|
||||||
const stringifySeconds = (seconds) =>
|
const stringifySeconds = (seconds) =>
|
||||||
new Date(seconds * 1000).toGMTString()
|
new Date(seconds * 1000).toISOString()
|
||||||
|
|
||||||
const hashKey = (key) =>
|
|
||||||
crypto.createHash('sha1').update(key).digest('hex')
|
|
||||||
|
|
||||||
// TODO: Copied from express-unpkg, use the same function
|
// TODO: Copied from express-unpkg, use the same function
|
||||||
const URLFormat = /^\/((?:@[^\/@]+\/)?[^\/@]+)(?:@([^\/]+))?(\/.*)?$/
|
const URLFormat = /^\/((?:@[^\/@]+\/)?[^\/@]+)(?:@([^\/]+))?(\/.*)?$/
|
||||||
|
@ -106,38 +86,44 @@ const getPackageName = (pathname) => {
|
||||||
|
|
||||||
const oneSecond = 1000
|
const oneSecond = 1000
|
||||||
const oneMinute = oneSecond * 60
|
const oneMinute = oneSecond * 60
|
||||||
const thirtyMinutes = oneMinute * 30
|
|
||||||
const oneHour = oneMinute * 60
|
const oneHour = oneMinute * 60
|
||||||
|
|
||||||
const computeLogChanges = (stream) =>
|
const computeLogChanges = (stream) =>
|
||||||
new Promise((resolve, reject) => {
|
new Promise((resolve, reject) => {
|
||||||
const changes = {}
|
const counters = {}
|
||||||
|
|
||||||
const incKey = (key, n = 1) =>
|
const incrKey = (key, by = 1) =>
|
||||||
changes[key] = (changes[key] || 0) + n
|
counters[key] = (counters[key] || 0) + by
|
||||||
|
|
||||||
|
const incrKeyMember = (key, member, by = 1) => {
|
||||||
|
counters[key] = counters[key] || {}
|
||||||
|
counters[key][member] = (counters[key][member] || 0) + by
|
||||||
|
}
|
||||||
|
|
||||||
stream
|
stream
|
||||||
.pipe(ndjson.parse())
|
.pipe(ndjson.parse())
|
||||||
.on('error', reject)
|
.on('error', reject)
|
||||||
.on('data', entry => {
|
.on('data', entry => {
|
||||||
const date = new Date(Math.round(entry.timestamp / 1000000))
|
const date = new Date(Math.round(entry.timestamp / 1000000))
|
||||||
const dayKey = `${date.getUTCFullYear()}/${date.getUTCMonth()}/${date.getUTCDate()}`
|
const dayKey = `${date.getUTCFullYear()}-${date.getUTCMonth()}-${date.getUTCDate()}`
|
||||||
const hourKey = `${dayKey}/${date.getUTCHours()}`
|
const hourKey = `${dayKey}-${date.getUTCHours()}`
|
||||||
|
const minuteKey = `${hourKey}-${date.getUTCMinutes()}`
|
||||||
|
|
||||||
// Q: How many requests do we receive per day/hour?
|
// Q: How many requests do we receive per day/hour/minute?
|
||||||
incKey(`requestsPerDay/${dayKey}`)
|
incrKey(`stats-requests-${dayKey}`)
|
||||||
incKey(`requestsPerHour/${hourKey}`)
|
incrKey(`stats-requests-${hourKey}`)
|
||||||
|
incrKey(`stats-requests-${minuteKey}`)
|
||||||
|
|
||||||
// Q: How many requests do we receive to edge/cache/origin per day/hour?
|
// Q: How many requests do we receive to edge/cache/origin per day/hour?
|
||||||
if (entry.origin) {
|
if (entry.origin) {
|
||||||
incKey(`originRequestsPerDay/${dayKey}`)
|
incrKey(`stats-originRequests-${dayKey}`)
|
||||||
incKey(`originRequestsPerHour/${hourKey}`)
|
incrKey(`stats-originRequests-${hourKey}`)
|
||||||
} else if (entry.cache) {
|
} else if (entry.cache) {
|
||||||
incKey(`cacheRequestsPerDay/${dayKey}`)
|
incrKey(`stats-cacheRequests-${dayKey}`)
|
||||||
incKey(`cacheRequestsPerHour/${hourKey}`)
|
incrKey(`stats-cacheRequests-${hourKey}`)
|
||||||
} else {
|
} else {
|
||||||
incKey(`edgeRequestsPerDay/${dayKey}`)
|
incrKey(`stats-edgeRequests-${dayKey}`)
|
||||||
incKey(`edgeRequestsPerHour/${hourKey}`)
|
incrKey(`stats-edgeRequests-${hourKey}`)
|
||||||
}
|
}
|
||||||
|
|
||||||
const clientRequest = entry.clientRequest
|
const clientRequest = entry.clientRequest
|
||||||
|
@ -146,68 +132,41 @@ const computeLogChanges = (stream) =>
|
||||||
const uri = clientRequest.uri
|
const uri = clientRequest.uri
|
||||||
const package = getPackageName(parseURL(uri).pathname)
|
const package = getPackageName(parseURL(uri).pathname)
|
||||||
|
|
||||||
if (package) {
|
if (package)
|
||||||
const key = `packageRequestsPerDay/${dayKey}/${hashKey(package)}`
|
incrKeyMember(`stats-packageRequests-${dayKey}`, package)
|
||||||
|
|
||||||
if (changes[key]) {
|
|
||||||
changes[key].requests += 1
|
|
||||||
} else {
|
|
||||||
changes[key] = { package, requests: 1 }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Q: How many requests per day do we receive via each protocol?
|
// Q: How many requests per day do we receive via each protocol?
|
||||||
const protocol = clientRequest.httpProtocol
|
const protocol = clientRequest.httpProtocol
|
||||||
|
|
||||||
if (protocol) {
|
if (protocol)
|
||||||
const key = `protocolRequestsPerDay/${dayKey}/${hashKey(protocol)}`
|
incrKeyMember(`stats-protocolRequests-${dayKey}`, protocol)
|
||||||
|
|
||||||
if (changes[key]) {
|
|
||||||
changes[key].requests += 1
|
|
||||||
} else {
|
|
||||||
changes[key] = { protocol, requests: 1 }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Q: How many requests per day do we receive from a hostname?
|
// Q: How many requests per day do we receive from a hostname?
|
||||||
const referer = clientRequest.referer
|
const referer = clientRequest.referer
|
||||||
const hostname = referer && parseURL(referer).hostname
|
const hostname = referer && parseURL(referer).hostname
|
||||||
|
|
||||||
if (hostname) {
|
if (hostname)
|
||||||
const key = `requestsPerDayAndRefererHostname/${dayKey}/${hashKey(hostname)}`
|
incrKeyMember(`stats-hostnameRequests-${dayKey}`, hostname)
|
||||||
|
|
||||||
if (changes[key]) {
|
|
||||||
changes[key].requests += 1
|
|
||||||
} else {
|
|
||||||
changes[key] = { hostname, requests: 1 }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
})
|
})
|
||||||
.on('end', () => {
|
.on('end', () => {
|
||||||
resolve(changes)
|
resolve(counters)
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
const processLogs = (stream) =>
|
const processLogs = (stream) =>
|
||||||
computeLogChanges(stream).then(changes => {
|
computeLogChanges(stream).then(counters => {
|
||||||
// Record the changes.
|
Object.keys(counters).forEach(key => {
|
||||||
Object.keys(changes).forEach(key => {
|
const value = counters[key]
|
||||||
const ref = db.ref(`logs/${key}`)
|
|
||||||
|
|
||||||
ref.transaction(value => {
|
if (typeof value === 'number') {
|
||||||
if (typeof changes[key].requests === 'number') {
|
// Simple counter.
|
||||||
// Nested value with a "requests" property.
|
db.incrby(key, value)
|
||||||
if (value && value.requests) {
|
} else {
|
||||||
value.requests += changes[key].requests
|
// Sorted set.
|
||||||
return value
|
Object.keys(value).forEach(member => {
|
||||||
} else {
|
db.zincrby(key, value[member], member)
|
||||||
return changes[key]
|
})
|
||||||
}
|
}
|
||||||
} else {
|
|
||||||
// Simple counter.
|
|
||||||
return (value || 0) + changes[key]
|
|
||||||
}
|
|
||||||
})
|
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -250,11 +209,11 @@ const ingestLogs = (zone, startSeconds, endSeconds) =>
|
||||||
})
|
})
|
||||||
|
|
||||||
const startZone = (zone) => {
|
const startZone = (zone) => {
|
||||||
const startSecondsRef = db.ref(`logs/nextStartSeconds/${zone.name.replace('.', '-')}`)
|
const startSecondsKey = `logsWorker-nextStartSeconds-${zone.name.replace('.', '-')}`
|
||||||
|
|
||||||
const takeATurn = () => {
|
const takeATurn = () => {
|
||||||
startSecondsRef.once('value', snapshot => {
|
db.get(startSecondsKey, (error, value) => {
|
||||||
let startSeconds = snapshot.val()
|
let startSeconds = value && parseInt(value, 10)
|
||||||
|
|
||||||
const now = Date.now()
|
const now = Date.now()
|
||||||
|
|
||||||
|
@ -283,13 +242,13 @@ const startZone = (zone) => {
|
||||||
// set of logs. This will help ensure that any congestion in the log
|
// set of logs. This will help ensure that any congestion in the log
|
||||||
// pipeline has passed and a full set of logs can be ingested.
|
// pipeline has passed and a full set of logs can be ingested.
|
||||||
// https://support.cloudflare.com/hc/en-us/articles/216672448-Enterprise-Log-Share-REST-API
|
// https://support.cloudflare.com/hc/en-us/articles/216672448-Enterprise-Log-Share-REST-API
|
||||||
const maxSeconds = toSeconds(now - thirtyMinutes)
|
const maxSeconds = toSeconds(now - (oneMinute * 30))
|
||||||
|
|
||||||
if (startSeconds < maxSeconds) {
|
if (startSeconds < maxSeconds) {
|
||||||
const endSeconds = startSeconds + LogWindowSeconds
|
const endSeconds = startSeconds + LogWindowSeconds
|
||||||
|
|
||||||
ingestLogs(zone, startSeconds, endSeconds).then(() => {
|
ingestLogs(zone, startSeconds, endSeconds).then(() => {
|
||||||
startSecondsRef.set(endSeconds)
|
db.set(startSecondsKey, endSeconds)
|
||||||
setTimeout(takeATurn)
|
setTimeout(takeATurn)
|
||||||
}, error => {
|
}, error => {
|
||||||
console.error(error.stack)
|
console.error(error.stack)
|
||||||
|
|
Loading…
Reference in New Issue