Node.js SQS to FastCGI worker/daemon/background process Welcome to this silly little utility program. It provides an intermediary link between an SQS message queue and a backend FastCGI system. It is very simple. All you
Alan Kent asked on Twitter What is your personal favorite HA distributed queue technology? RabbitMQ? Kafka? Redis? ActiveMQ? Why? — Alan Kent (@akent99) February 4, 2015 You may know that I am partial to
ECommerce is a small thing, right? Nobody’s doing it and it’s so simple that everyone who does it is doing it right. When that Cyber Monday hits, nobody panics; sites stay up, they’re able to handle the load and nobody gets yelled at, right?
OK, maybe 20 years ago.
I just had an epiphany. I’ve talked about pre-caching content before and the benefits thereof before. But this is the first time I realized not only that there are benefits, but that doing it is BETTER than caching inline. Let me sum up… no, there is to much. Let me explain.
Typically caching is done like this (stolen from the ZF caching docs):
So there I was, looking at some other websites out there (because I think my site design sucks. Thanks, me). One of the things that virtually no blogs do is promote specific content. In other words, highlight content that is most popular over a certain time frame. So I was thinking to myself, how would I do that? One option would be to have a database table that could record each click. That, however, is boring and requires changes to my DB schema (evil!). What I want to do is take my most popular pages of the last week and highlight them at the top of the web site.
Working with an off-the-shelf shopping cart usually requires a little bit of patience. Scaling an e-commerce site does have its share of problems. There is a LOT of interactivity that needs to be implemented. This can be things along the lines of generating targeted ads, sending email or charging a credit card.
With the web being what it is today there can be a lot of times when you want to aggregate data from many different sources and bring them together in a single page. I have not done much of that on my site simply because that means that I then need to learn a bunch of different API’s. However, since Youtube is the #2 search engine I figured that it might not be a bad idea to aggregate some of my YouTube content on my page automatically. I don’t necessarily want to do a blog post about each individual video I post, but I wanted there to be some place where I could just list them out.
In our previous installment we looked at setting our backend up so it could automatically retrieve the bit.ly URL for a given URL and store it as part of the data for a given instance of a Content model. What we’re going to do this time is take a look at the front end components.
Sometimes I find that doing things backwards can actually make things a little more clear. That way you can see the end result and then, as you work backwards, see how all the pieces work together.