Be sure to check out <a href="https://github.com/snowplow/snowplow" rel="nofollow">https://github.com/snowplow/snowplow</a> before rolling out your own custom analytics solution..
Very interesting. You say "Within an hour or so each GET shows up in the logs for the S3 bucket wherever we’re sending the logs for our S3 example.bucket." Is this frequency configurable? I also see at the link below that the logs are best effort and they may drop messages. How is this for you in practice?<p><a href="http://docs.aws.amazon.com/AmazonS3/latest/dev/ServerLogs.html" rel="nofollow">http://docs.aws.amazon.com/AmazonS3/latest/dev/ServerLogs.ht...</a>
That's really cool. You're saving money in 2 places:<p>1) not paying for SQS, Kinesis, or some other kind of queue to store the events.<p>2) not paying for a backend queue processor job to read the events from the queue and write them to S3/Redshift.<p>You still need a backend processor to do the Redshift copy commands, but this is much lighter weight and you'd never need to scale this process.<p>You'd also be paying some additional money for the S3 logging but probably pretty minimal.<p>Very cool!