InfoQ Homepage QCon London 2024 Content on InfoQ
-
Stateful Cloud Services at Neon Navigating Design Decisions and Trade-Offs: Q&A with John Spray
At QCon London, John Spray, a storage engineering lead @neon.tech, discussed the often-overlooked complexities of stateful cloud service design, using Neon Serverless Postgres as a case study. His session was part of the Cloud-Native Engineering track on the first day of the conference, and InfoQ carried out an interview.
-
QCon London: Scaling Microservices Architecture and Technology Organization at Trainline
During the recent QCon London conference, Trainline’s CTO spoke about the evolution of the company’s system architecture and organizational structure over the last five years. The company had to adapt to market changes and growing customer expectations by improving the performance and reliability of its technology platform.
-
Fix Your SDLC before Adopting Gen AI in Your Organisation: Bannon’s Call to Action at QCon London
During her keynote at QCon London, Tracy Bannon, architect and researcher at MITRE, argued that AI will be able to enhance the software development lifecycle, though currently it’s at the “code completion” rather than “code generation” phase. Throughout her presentation, she continuously stresses the importance of keeping humans in the loop and fixing your company’s SDLC before embracing AI.
-
QCon London: Lessons Learned from Building LinkedIn’s AI/ML Data Platform
At the QCon London 2024 conference, Félix GV from LinkedIn discussed the AI/ML platform powering the company’s products. He specifically delved into Venice DB, the NoSQL data store used for feature persistence. The presenter shared the lessons learned from evolving and operating the platform, including cluster management and library versioning.
-
Building a Platform to Gain an Unexpected Competitive Advantage: Ranbir Chawla at QCon London
During his QCon London presentation, Ranbir Chawla presented the journey his team took from moving from an “architectural perfect storm” and a highly manual operational system to a product company with a modern event-based architecture that can be released in < 1 hour. The company now focuses on providing real business outcomes to its stakeholders, and ensuring developers find joy in their work.
-
Borderless Cloud at QCon London: Q&A with Adora Nwodo
At QCon London, Adora Nkowno, senior software engineer at NexaScale, discussed the complexities of seamlessly integrating multiple clouds into application architecture, deployment processes, and CI/CD pipelines. Her session was part of the Cloud-Native Engineering track on the first day of the conference, and InfoQ did an interview.
-
QCon London: gRPC Migration Automation at LinkedIn
At QCon London 2024, Karthik Ramgopal and Min Chen described how AI helped LinkedIn change the remote procedure calls (RPC) protocol for 50,000 production endpoints from Rest.li to Google's gRPC. A planned 2-3 year manual migration turned into an AI-supported migration lasting 2-3 quarters. It changed 20 million lines of code across 2000 services – without business interruption.
-
QCon London: Mastering Long-Running Processes in Modern Architectures
At QCon London 2024, Bernd Ruecker recommended implementing long-running tasks asynchronously with a process-orchestration platform. Such a platform provides better service boundaries and efficiencies and reduces accidental system complexity and risk. Organizing the platform centrally in an organization eases orchestration adoption by applications.
-
QCon London: The Art, Science and Psychology of Decision-Making
At QCon London 2024, Hannes Ricklefs, head of architecture at the BBC, gave a well-received talk on decision making. Ricklefs summarised the key reasons behind applying art, science and psychology to the discipline of decision-making, focusing on appropriate methodologies to use and the effects of biases on our ability to make good decisions in both a personal and business context.
-
QCon London: How Duolingo Sent 4 Million Push Notifications in 6 Seconds During the Super Bowl Break
As part of the Super Bowl marketing campaign, Duolingo sent out 4 million mobile push notifications when the company’s five-second ad aired during the commercial break. At QCon London, Doulingo’s engineers presented the asynchronous AWS architecture responsible for broadcasting messages to millions of users across seven US cities.
-
QCon London: Efficient Serverless Development
At QCon London, Yan Cui, a serverless advocate at Lumigo, shared patterns for effective local development with AWS serverless technologies. The focus areas were testing approaches, deployment practices, and application environments.
-
Large Language Models for Code by Loubna Ben Allal at QCon London
At QCon London, Loubna Ben Allal discussed Large Language Models (LLMs) for code. She discussed the lifecycle of code completion models, which consists of pre-training on vast codebases and finetuning and continuous adaptation. She specifically discussed open-source models, which are powered by platforms like Hugging Face.
-
QCon London: Meta Used Monolithic Architecture to Ship Threads in Only Five Months
Zahan Malkani talked during QCon London 2024 about Meta’s journey from identifying the opportunity in the market to shipping the Threads application only five months later. The company leveraged Instagram's existing monolithic architecture and quickly iterated to create a new text-first microblogging service in record time.
-
Efficient DevSecOps Workflows with a Little Help from AI: Q&A with Michael Friedrich
At QCon London, Michael Friedrich, senior developer advocate at GitLab, discussed how AI can help in DevSecOps workflows. His session was part of the Cloud-Native Engineering track on the first day of the conference. InfoQ interviewed Friedrich after the session.
-
Navigating LLM Deployment: Tips, Tricks and Techniques by Meryem Arik at QCon London
At QCon London, Meryem Arik discussed deploying Large Language Models (LLMs). While initial proofs of concept benefit from hosted solutions, scaling demands self-hosting to cut costs, enhance performance with tailored models, and meet privacy and security requirements. She emphasized understanding deployment limits, quantization for efficiency, and optimizing inference to fully use GPU resources.