{Speakers}
No unicorns, no caticorns, just software development
Heidi Howard
Senior Researcher at Microsoft
Heidi Howard
Senior Researcher at Microsoft
I am a Senior Researcher in the Confidential Computing group at Microsoft Research Cambridge. My research sits at the intersection between the theory and practice of distributed computing, with a focus on developing resilient and trustworthy distributed computer systems. Previously, I was a Research Fellow in Computer Science at Cambridge University’s Trinity Hall, an Affiliated/Visiting Researcher at VMware Research, and an Affiliated Lecturer at Cambridge University’s Department of Computer Science and Technology. I received my Ph.D. from Cambridge University in 2019 for my research on Distributed Consensus. I am probably best known for my work on the Paxos algorithm, and in particular, the invention of Flexible Paxos.
Confidential Consortium Framework: Building Secure Multiparty Applications in the Cloud (Without Handing Over the Keys to the Kingdom!)
In the pre-cloud era, computer systems were operated by the organizations which depended upon them. This on-premises approach gave organizations great power over their systems, however, “with great power comes great responsibility” and organizations were left with the ongoing burden of deploying and managing their own infrastructure. Today, Cloud computing has removed much of the responsibility of deploying systems, however, it has also removed much of the power that organizations once had. Organizations must place their trust in the cloud to secure the confidentiality and integrity of their data.
In this talk, I'll consider whether it is possible to regain control over data in the cloud (great power with none of the responsibility) and even enable multiple untrusted parties to compute together on untrusted infrastructure. I’ll introduce the Confidential Consortium Framework (aka CCF), an open-source framework for building a new category of secure multiparty applications with confidentiality, integrity protection, and high availability. CCF utilizes hardware-based trusted execution environments for remotely verifiable confidentiality and code integrity, backed by an auditable and immutable distributed ledger for data integrity and high availability. CCF even enables application developers to bring both their own application logic and a custom multi-party governance model, in the form of a programmable constitution. By the conclusion of this talk, I hope to have convinced you that distributing systems does not necessarily mean distributing trust in the era of confidential computing in the cloud. You can learn more about CCF today at: https://ccf.dev/
Andy Pavlo
Associate Professor of Databaseology at Carnegie Mellon University
Andy Pavlo
Associate Professor of Databaseology at Carnegie Mellon University
Andy Pavlo is an Associate Professor (Indefinite Tenure) of Databaseology in the Computer Science Department at Carnegie Mellon University. He is also the co-founder of the OtterTune automated database optimization start-up (https://ottertune.com). He is from the streets.
Why Machine Learning for Automatically Optimizing Databases Doesn't Work
Database management systems (DBMSs) are complex software that requires sophisticated tuning to work efficiently for a given workload and operating environment. Such tuning requires considerable effort from experienced administrators, which is not scalable for large DBMS fleets. This problem has led to research on using machine learning (ML) to devise strategies to optimize DBMS configurations for any application, including automatic physical database design, knob configuration, and query tuning. Despite the many academic papers that tout the benefits of using ML to optimize databases, there have been only a few major success stories in industry in the last decade.
In this talk, I discuss the challenges of using ML-enhanced tuning methods to optimize databases. I will address specific assumptions that researchers make about production database environments that are incorrect and identify why ML is not always the best solution to solving real-world database problems. As part of this, I will discuss state-of-the-art academic research and real-world tuning implementations.
Nadieh Bremer
Data Visualization Artist at Visual Cinnamon
Nadieh Bremer
Data Visualization Artist at Visual Cinnamon
Nadieh Bremer is a data visualization artist that once graduated as an Astronomer, started working as a data scientist before finding her true passion in the visualization of data. As 2017's "Best Individual" in the Information is Beautiful Awards, and co-writer of "Data Sketches", she focuses on visuals that are uniquely crafted for each specific dataset, often using large and complex datasets while employing vibrant color palettes. She's made visualizations and art for companies such as Google News Lab, Sony Music, UNICEF, the New York Times and UNESCO.
Visualizing Connections
Connections are a part of us, of the world. From the connections between people, between cultures, within language, and more. In these days when more data is collected daily than we could ever hope to explore, the variety in connections being gathered is opening up the possibility to visualize these (often complex) networks. During this talk, Nadieh will take you through the design process of several of her (interactive) data visualization works, from personal projects to client work. The common thread they all share, is that they all reveal connections, but all differently. From a family tree of 3000 people connected to the European Royal houses, to those existing between our Intangible Cultural Heritage created for UNESCO, to connections we have drawn in the night skies, something with cats and dogs, and more. Revealing that all types of connections are unique and revealing the intricacies that lie within them requires a creative, iterative and custom approach.
Colin Breck
Cloud Platforms Lead at Tesla
Colin Breck
Cloud Platforms Lead at Tesla
Colin leads the cloud platforms organization for Tesla Energy developing real-time services and critical infrastructure for power generation, battery storage, vehicle charging, and grid services. Over the past six years, he has seen these platforms grow from their infancy to become the largest and most integrated platforms for distributed, renewable energy in the world. Previously, Colin worked on the PI System at OSIsoft, a time-series platform for industrial monitoring and automation.
Kubernetes Probes: How to Avoid Shooting Yourself in the Foot
Kubernetes liveness and readiness probes can be used to make a service more robust and more resilient, by reducing operational issues and improving the quality of service. However, if these probes are not implemented carefully, they can severely degrade the overall operation of a service, to a point where you would be better off without them. I will explore how to avoid making service reliability worse when implementing Kubernetes liveness and readiness probes by learning from production incidents.
Santiago Valdarrama
Machine Learning Engineer at his own
Santiago Valdarrama
Machine Learning Engineer at his own
Santiago is a Machine Learning Engineer instructor. He has a Master's in Machine Learning from the Georgia Institute of Technology and two decades of experience building software for some of the largest companies in the world. He co-founded bnomial.com, where he publishes daily Machine Learning questions and competitions.
We made our robots talk!
Is there a better way to communicate with a robot than using natural language? In this talk, you'll learn how we integrated Boston Dynamic's Spot robot with ChatGPT. You'll learn how we use computer vision to solve impossible problems and how ChatGPT 10x'd Spot's capabilities overnight.
Paul Nordstrom
Principal Engineer at Temporal Technologies
Paul Nordstrom
Principal Engineer at Temporal Technologies
My career has followed a bit of a winding road. I started out in the financial markets: Risk Management for a major financial market clearing firm; partner in an Algorithmic Trading company; founder/owner of a hedge fund; president of a German Bank. Then back to my roots as a software engineer developing low-latency, high-volume proprietary trading systems. I moved over to Web 1.0 as Amazon's first Sr. Principal engineer where I led a team building to replace the existing (monolithic) website architecture with one of the earliest Services-Oriented Architectures for a large-scale website. Also designed/wrote Amazon's RPC framework, standard service application framework, source code repository, build system, and a few other things. On to Google, where I designed and led a team in the building of Millwheel, one of the earliest high-scale continuous computation frameworks, and finally now a Principal Engineer at Temporal Technologies, where we are building the foundation of a comprehensive set of tools for Cloud Applications.
Software Organisms
The past and future evolution of software development as we know it.
This talk doesn't try to "prove" anything, except that I have a long history in this industry :)
It's an argument-by-analogy (and therefore immediately suspect, yes?) about how we got to where we are (hint: evolution) and where we are going (hint: more evolution) writing software design in the large. The implication is that the Cloud revolution is still just beginning. If that's so, how will we be building Cloud systems 10 years from now?
This talk takes a stab at answering that question while trying to be "concurrently" interesting and entertaining.
Stefan Tilkov
Founder at INNOQ
Software Architecture, Processes, Organization — and Humans
In this talk, we’ll look at software architecture, team organization and the interplay between technology and humans. We’ll address some of the patterns and anti-patterns that can be observed when organizations try to evolve towards more decentralization and autonomy, and take a look at some strategies to ensure technology supports and enable the desired outcome. We’ll also mention data mesh as it’s 2023 and we have to.
Michał Maślanka
Software Developer at Redpanda
Michał Maślanka
Software Developer at Redpanda
Michal has 10+ years of experience in software engineering across different industries, focusing on distributed systems. He joined Redpanda in 2019 where he is one of the primary contributors to Redpanda core. Michal is currently responsible for Redpanda's Raft implementation and cluster orchestration bits.
What can we learn from Control Theory ? - a deep insight into Redpanda backlog controller
Redpanda is a Kafka®-compatible streaming data platform that not only handles user writes and reads, but also needs to maintain a tremendous number of background processes—including log compaction, object store uploads, data and leadership balancing. The challenge is that these background tasks compete with the main workload for finite resources, like CPU, I/O bandwidth and memory.
Some of these background tasks share a common property: a growing backlog.
Keeping the backlog size small with minimal influence on write/read path latency requires tuning the number of shares given to the background task.
This talk walks you through our challenges with scheduling background tasks in Redpanda, and brings you behind the scenes to learn how we leveraged the PID controller (an idea borrowed from industrial applications) to manage the amount of shares given for background tasks with a growing backlog.
Dr. Roland Kuhn
CTO at Actyx
Dr. Roland Kuhn
CTO at Actyx
Roland is CTO and co-founder of Actyx, author of Reactive Design Patterns, a co-author of the Reactive Manifesto, and teacher of the edX course Programming Reactive Systems. Previously he led the Akka project at Lightbend. He also holds a Dr. rer. nat. in particle physics from TU München and has worked in the space industry. He spends most of his life in the central European timezone.
"All Matters Rust" Unconfakerence
Based on the core concepts of unconferences and open spaces we have created a new disruptive, scalable, fully decentralised, secure, and resilient session called unconfakerence.
While unconferences are supposed to be participant-driven meetings without an agenda or topic to be discussed, in the case of this unconfakerence there will be a predefined topic about Rust (aka RUST for some C++ programmers) on which different suggestions and subtopics will be raised.
The organisation will be led by Dr Roland Kuhn with the help of Arno Schots, and Gonzalo G. Jimenez, members of the JOTB organizing committee.
The Law of 2 feet prevails
If at any time during the session, a person feels that they are not learning or contributing, they have a responsibility to get up and move, use both feet to move on to something interesting.
Peer-To-Peer Workflows: Modelling, Validation, and Execution
When you start modelling a workflow, chances are that you sketch some boxes on a whiteboard with arrows between them. Each of the arrows means that some participant does something to advance the workflow (or loop back to an earlier state). This very intuitive approach turns out to be appropriate not only for BPMN engines but also for fully decentralised systems, where participants interact in a peer-to-peer mesh network fashion. This talk presents the Actyx machine-runner library, including the distributed systems theory powering the machine-check companion tool. The latter allows you to make sure that the machines you wrote for each participant actually implement the workflow correctly, and that following the workflow provides eventual consensus — entirely free from central components or coordination.
Pedro Holanda
COO at DuckDB
Pedro Holanda
COO at DuckDB
Pedro Holanda is a computer scientist with a background in database architectures. He completed his Ph.D. at CWI in Amsterdam, where he specialized in indexing for interactive data analysis. He is a prominent contributor to the open-source database management system, DuckDB. Currently, he is the COO of DuckDB Labs, a company that provides services and support for DuckDB.
In-Process Analytical Data Management with DuckDB
Analytical data management systems have been hard to use, expensive and far removed from the actual processes. However, by revamping these systems to integrate with the application process, data transfer, deployment, and management can be significantly streamlined. This innovative approach opens up a plethora of new possibilities, including edge OLAP, running SQL queries in lambdas, and analyzing Big Data on laptops.
Enter DuckDB, a novel analytical data management system that has been specifically designed for in-process use cases. DuckDB is SQL-enabled, easily integrated as a library, and boasts cutting-edge query processing techniques, such as vectorized execution and lightweight compression. It is also free and open-source software, distributed under the permissive MIT license. In my presentation, I will delve into the reasoning and design principles that underpin DuckDB, and offer a comprehensive overview of its inner workings.
Simon Ritter
Deputy CTO at Azul Systems
Simon Ritter
Deputy CTO at Azul Systems
Simon Ritter is the Deputy CTO of Azul Systems. Simon joined Sun Microsystems in 1996 and spent time working in both Java development and consultancy. He has been presenting Java technologies to developers since 1999 focusing on the core Java platform as well as client and embedded applications. At Azul, he continues to help people understand Java and AzulÕs JVM products.
Simon is a Java Champion and two time recipient of the JavaOne Rockstar award. In addition, he represents Azul on the JCP Executive Committee, the OpenJDK Vulnerability Group as well as the JSR Expert Group since Java SE 9.
The Cloud Native Compiler: JIT-as-a-Service
Adaptive, just in time (JIT) compilation provides a massive performance improvement to JVM-based applications compared to only using an interpreter. The downside of this is that applications have to compile frequently used methods as the application is running. This can lead to reduced throughput and slower response times. Another drawback is that each time an application is started, it must perform the same analysis to identify hot spot methods and compile them. When running an application in the cloud, the elastic nature of resources provides the ability to change and improve the dynamics of how the JIT compiler works. In this session, we'll look at Azul's work to move the JIT compiler into a centralized service that can be shared by many JVMs. This provides many advantages, such as caching compiled code for instant delivery when restarting the same application or spinning up new instances of the same service. In addition, it removes the workload from individual JVMs, allowing them to deliver more transactions per second of application work. Finally, there is the opportunity to apply considerably more compute resources to enable complex optimizations to be used that wouldn't be practical in a single JVM.
Rashmi Nagpal
Software Engineer at her own
Rashmi Nagpal
Software Engineer at her own
Rashmi is a Software Engineer with a passion for building products in AI/ML. In her almost 4 years career in tech, she’s brought products to life at pre-seed startups, scaled teams and software at hypergrowth unicorns, and shipped redesigns and features used by millions at established giants. When she's not coding, capturing cosmos using her telescope, or playing board games with friends, you can find Rashmi playing with her maltese breed pet dog, Fluffy!
Unearth The Black-Box : Building Fair, Accountable and Trustworthy ML Systems
Have you ever wondered why 87% of machine learning models never make it to production? Who must be held responsible if a machine learning algorithm discriminates or shows bias? Are the decisions taken by these models trustworthy? In this talk, let’s unravel the answers to such complex questions!
Machine learning has had a significant impact in many areas, including medicine, entertainment, security, and education, but its use can also result in increased cognitive dependence on technology and ethical concerns such as bias. Therefore, it is crucial to address these issues by reducing the impact of human biases and creating trustworthy, reliable, and understandable machine learning systems.
The key takeaways of my talk would likely include the importance of understanding and interpreting the decision-making processes of machine learning models, as well as the need to ensure that these models are fair, accountable, and trustworthy in their predictions and actions. Additionally, the talk may highlight the challenges of building interpretable models and the importance of evaluating and testing models for bias, as well as the need for transparency and accountability in the development and deployment of machine learning systems.
Carlos Baquero
Professor at Universidad do Porto
Carlos Baquero
Professor at Universidad do Porto
Carlos Baquero is a Professor in the Department of Informatics Engineering within FEUP, and area coordinator at the High Assurance Laboratory (HASLab) within INESC TEC. From 1994 till mid-2021 he was affiliated with the Informatics Department, Universidade do Minho, where he concluded his PhD (2000) and Habilitation/Agregação (2018). He currently teaches courses in Operating Systems and in Large Scale Distributed Systems. Research interests cover data management in eventual consistent settings, distributed data aggregation and causality tracking. He worked in the development of data summary mechanisms such as Scalable Bloom Filters, causality tracking for dynamic settings with Interval Tree Clocks and Dotted Version Vectors and predictable eventual consistency with Conflict-Free Replicated Data Types. Most of this research has been applied in industry, namely in the Riak distributed database, Redis CRDBs, Akka distributed data and Microsoft Azure Cosmos DB.
The Legacy of Peer-to-Peer Systems
This talk introduces some of the history and evolution of Peer-to-Peer (P2P) systems in the last 20 years. Although the novelty of the concept has faded, it did still provide many contributions to the design of distributed systems that are still relevant nowadays. The talk address the connections to filesystem research from the 80s, the evolution of DHTs, and its influence on the NoSQL DBs and Blockchains.
Jesús María Villar
Senior Software Engineer at Celonis
Jesús María Villar
Senior Software Engineer at Celonis
Jesus is a passionate developer with 18 years of experience and sound knowledge of the software development lifecycle. While his current focus is Java Backend development, in the past Jesus has worked in very different stacks: PHP, Javascript, Typescript, Python, Dart, C++, and even COBOL! Considered one of the foremost XP advocates by the Spanish Community, he is now working at Celonis where he is starting to dive into the process mining and AI worlds. Want to learn more? Besides coding and best practices, Jesus loves Extreme Sports and Punk Rock music.
Can Engineers Save the Planet?
Traditional vs Process Mining
Process Mining has become a major enabler in successfully bringing operational excellence to business processes. The journey to implementing Process Mining will be similar but the outcome (insights and actions) will be unique to each company.
Traditional approaches fail to understand the real-life complexity of processes and also struggle to provide complete insights given the vast amounts of data that are now available. By contrast, Process Mining offers a data-driven and more objective and holistic approach to understanding business processes. As a result, Process Mining has come to dominate a large majority of operational excellence, automation and digitalization ambitions within the industry.
Basically Process Mining will reveal the truth of your data, because the data don’t lie
Why Process Mining?
Process Mining is the leading new technology when it comes to talking about algorithmic businesses - in other words, businesses that use algorithms and large amounts of real-time data to create business value. Global Process Mining Software Market is valued approximately at USD 322.02 Million in 2020 and is anticipated to grow more than 50.1% by 2027.
How do we plan to save the world?
At Celonis, we’ve been applying Process Mining for over a decade and in this session, I will share the principles, the main algorithms and our approach that have helped us to scale and evolve our platform to support hundreds of companies to optimize their processes and drive business success while helping to create a better and more sustainable world.
Jônatas Davi Paganini
Developer Advocate at Timescale
How to model a Time Series database with TimescaleDB
Storing massive time series data is always a challenge. Modeling it brings several doubts on how to balance heavy data throughput and still easy to query the data. Join the talk to learn the pros and cons of wide versus narrow structures for storing time series data.
This talk is an overview of time-series database storage and how to create a structure that can help you to maintain your systems in the long term.
Timescaledb is a Postgresql extension that brings the time-series superpowers to plain sql. This talk will discuss wide and narrow models and how to leverage the extension features in a common DevOps scenario.
Federico Fregosi
VP of Engineering at Contino
Federico Fregosi
VP of Engineering at Contino
Experienced engineering leader with a strong interest in distributed, highly-scalable and cloud-based systems. Currently at Contino as VP of Engineering where he oversees a team of 150+ highly talented, intelligent and thought-provoking technical engineers from a range of disciplines and backgrounds. Worked for years in Technical Leadership roles, focused on infrastructure management and Cloud on the 3 major public cloud providers. Experienced in the finance sector, specifically on high-performance payment platforms and systems compliance. Regular speaker at conferences and meetups. Federico holds an MSc in Software Engineering from City University London Experienced engineering leader with a strong interest in distributed, highly-scalable and cloud-based systems.
Containers in the cloud - State of the Art in 2023
In only a few years, the number of options available to run containers in the Cloud has literally exploded. Each provider now offers tens of “slightly different” services, each with its own minor trade-offs. Furthermore, running your applications in 2023 is definitely not like doing it in 2019: some of the new serverless options offer unique value propositions that shouldn’t be missed. It’s easy to get overwhelmed! This talk will categorize the various options available on AWS, Azure & GCP, focusing on what is state-of-the-art in 2023. We’ll look at Kubernetes and its evolution. Finally, we’ll explain the trade-offs between different categories from a technical and organizational standpoint. We’ll then do a deep dive with a demo on some of the new services that have been recently launched and that are quickly evolving to change the game: GCP Cloud Run, Azure Container Apps, and AWS Fargate.
Kubernetes & The Myth of Cloud Portability
Engineers at small and large companies are deploying their applications using managed Kubernetes services in the cloud because they are convinced it will enable them to migrate their workloads across providers easily. However, in reality, most container platforms are not feature-complete. So, the move comes with a series of external concerns that can make it problematic: e.g. monitoring, identity management, data gravity, deployment pipelines, dependent services, and many more!
In this talk, we’ll follow the migration journey of a demo application from AWS to GCP and we will identify a series of practical software development and system architecture best practices that you can follow to increase the level of portability of your Kube application.
Furthermore, we will consider configuration and components in the Kubernetes ecosystem that you can deploy to significantly simplify any platform lock-in constraints.
Takeaways
- Understand the portability limitations for microservices deployed on Kubernetes in the cloud
- Best practices to develop and configure your application to increase its chances of portability
- Defining a framework that can help assess the complexity of moving a Kubernetes application from one cloud provider to the other
Kaviraj Kanagaraj
Senior Software Engineer at Grafana Labs
Kaviraj Kanagaraj
Senior Software Engineer at Grafana Labs
Kaviraj is a Senior Software Engineer at Grafana Labs, working on highly scalable storage for Logs, One of the core maintainers of Grafana Loki Open Source project.
In his previous life, writing and maintaining C++ modules for L2/L3 traffic management for telecom backend systems. Moved to web backend and distributed systems later in his career writing bits of Python and Go. Big UNIX fanboy and passionate about OS abstractions and internals.
Getting started with Grafana Loki, a modern logs database
Loki is a modern log database that has different design tradeoffs (index only metadata) compared to traditional logs databases(index everything). This makes Loki easy to use and operate at Peta Byte scale. It can also use cheaper cloud storages (e.g S3) as persistent storage for indexes and chunks.
First, we try to understand why Loki? Why do we even need a modern logs database? The data model of Grafana Loki is different from traditional logs databases. We will explore why that difference matters when handling logs at a huge scale(Peta Byte) and how that makes it easy for both Loki operators and Loki users. We will then explore how to use Loki. We scrape logs from different targets and send them to Loki, then we use LogQL (powerful query language for logs, inspired from PromQL) to get visibility of your logs instead of just distributed grep. We also explore a few best practices on logging patterns that we use internally and how it helps to effectively investigate SLOW query/endpoints of your application and services.
Sometimes logs may not give you the complete picture, We explore how Loki integrates well with metrics and traces to enhance the observability experience of the users. Finally, we touch on some of the new and upcoming features of Loki.
Marlene Mhangami
Developer Advocate at Voltron Data
Marlene Mhangami
Developer Advocate at Voltron Data
Marlene is a Zimbabwean software engineer, developer advocate and explorer. She is a previous director and vice-chair for the Python Software Foundation and is currently serving as the vice-chair of the Association for Computing Machinery practitioner board. In 2017, she co-founded ZimboPy, a non-profit organization that gives Zimbabwean young women access to resources in the field of technology. She is also the previous chair of PyCon Africa and is an advocate for women in tech on the continent. Professionally, Marlene is currently working as a Developer Advocate at Voltron Data.
Elephants, ibises and a more Pythonic way to work with databases
In this talk, I will be sharing about Ibis, a software package that provides a more Pythonic way of interacting with multiple database engines. In my own adventures living in Zimbabwe, I’ve always encountered ibises (the bird versions) perched on top of elephants. If you’ve never seen an elephant in real life I can confirm that they are huge, complex creatures. The image of a small bird sitting on top of a large elephant serves as a metaphor for how ibis (the package) provides a less complex, more performant way for Pythonistas to interact with multiple big data engines.
I'll use the metaphor of elephants and ibises to show how this package can make a data workflow more Pythonic. The Zen of Python lets us know that simple is better than complex. The bigger and more complex your data, the more of an argument there is to use Ibis. Raw SQL can be quite difficult to maintain when your queries are very complex. For Python programmers, Ibis offers a way to write SQL in Python that allows for unit-testing, composability, and abstraction over specific query engines (e.g.BigQuery)! You can carry out joins, filters, and other operations on your data in a familiar, Pandas-like syntax. Overall, using Ibis simplifies your workflows, makes you more productive, and keeps your code readable
Javier Ramírez
Developer Advocate at Quest DB
Javier Ramírez
Developer Advocate at Quest DB
As a Developer Advocate at QuestDB, Javier helps developers make the most of their (fast) data, He makes sure the core team behind QuestDB listens to absolutely every piece of feedback he gets, and he facilitates collaboration in their open-source repository.
Javier loves data storage, big and small. He has extensive experience with SQL, NoSQL, graph, in-memory databases, Big Data, and Machine Learning. He likes distributed, scalable, always-on systems.
Ingesting over a million rows per second on a single database instance
How would you build a database to support sustained ingestion of several hundreds of thousands of rows per second while running near real-time queries on top?
In this session, I will go over some of the technical decisions and trade-offs we applied when building QuestDB, an open-source time-series database developed mainly in JAVA, and how we can achieve over a million row writes per second on a single instance without blocking or slowing down the reads. There will be code and demos, of course.
We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read replicas, or faster batch ingestion.
Xiaoman Dong
Software Engineer at StarTree
Xiaoman Dong
Software Engineer at StarTree
Xiaoman Dong has devoted his past 10+ years working in the streaming analytics and database domain, building data infrastructure, scalable distributed systems, and low latency queries over large datasets. During his work in StarTree, and Uber, he has designed, led, built, and operated several large-scale business-critical solutions based on open-source software like Apache Kafka, Apache Pinot, Apache Flink, and Kubernetes. While working in Stripe, he has also built and run the world’s largest single Pinot cluster with around 1 trillion rows and 1 PB in size.
Xiaoman is also an advocate of Big Data and distributed OLAP systems. He has been actively speaking at large tech conferences Kafka Summit, Flink Forward, and ApacheCon in recent years.
Kubernetes Clusters At Scale: Managing Hundreds Apache Pinot Kubernetes Clusters Inside Each End User’s Own Cloud Infrastructure
How to efficiently build and manage hundreds of Kubernetes Clusters that serve modern online analytics databases, for different customers? To add to the challenge, what if customers need to run their own clusters inside their own private clouds? We are sharing our system design that solves it.
How to provide fully managed online analytics databases like Pinot to hundreds of customers, while those Pinot clusters are running in each customer’s own private virtual cloud? The answer is by combining the power of Kubernetes with our automated scalable architecture that can fully manage a fleet of Kubernetes clusters.
When companies consider using SaaS (Software as a Service) products, they are often held back by challenges like security considerations and storage compliance regulations. Those concerns often require that the data stays within the same virtual cloud owned by the company. And it makes managed solutions very hard for companies to implement.
In StarTree we have built a modern data infrastructure based on Kubernetes so companies can keep their data inside their own infrastructure, and at the same time get the benefits of using a fully managed Apache Pinot cluster deployed in the customer’s cloud environment.
We have designed a scalable system based on Kubernetes that enables remote creation, maintenance, and monitoring of hundreds of Kubernetes clusters from different companies. This enabled us to scale quickly from a handful of deployments to over 100+ Pinot clusters in a short time span with just 10+ engineers.
Valerio Bartolini
Senior Software Engineer at Red Hat
Valerio Bartolini
Senior Software Engineer at Red Hat
Valerio is a Senior Software Engineer at Red Hat with 10 years of experience, developing software in enterprise environments as well as startup. He holds a Master's Degree in Computer Engineering and has lived in 4 countries. Open source and sports enthusiast.
Building multi-cloud applications with Skupper
Building a multi-cloud network requires complex VPN configurations and policies. How can we enable secure communication across Kubernetes clusters without these problems? By making Skupper we have been facing these challenges and in this discussion, I want to present what problems we solve and what architectural decisions we use. Finally, a demo to show an actual use case.
Vanessa Formicola
Engineering Manager at Flo
Vanessa Formicola
Engineering Manager at Flo
Engineering Manager at Flo, 10+ years of industry experience (Microsoft/ThoughtWorks) and 3+ years in leadership roles. Experienced with high scale backend systems, legacy modernisation and infrastructure as code across multiple domains and tech stacks. Primarily focused on transformation of software systems and enablement of teams with varied backgrounds. Experienced working with distributed global teams (multi-tz). Creator of knowledge sharing/D&I communities, public speaker and social change advocate.
Tech Leading 101: the Good, the Bad and the Ugly
Tech Leading is a hard job. You need to be an architect, carer, leader, servant, software engineer, business analyst, DevOps engineer, diplomat and a lot more all at the same time… What it means to be the tech lead of a team will be explored, and its joys and responsibilities. More importantly, we will navigate the complex balance of responsibilities and the wide variety of skills needed. Real-life experiences will be shared and lots of advice on becoming the best tech lead you can be. Whether you would like to become one, you currently are doing the job or you are leading tech leads this talk might be interesting for you.
Juanjo Madrigal
Software Engineer at Devo
Juanjo Madrigal
Software Engineer at Devo
Juanjo Madrigal is a software engineer at Devo. He moved from pure maths (MSc in Advanced Mathematics, Geometry & Topology, UCM) to programming, passing through fields like Computer Vision, Deep Learning and AI. During this journey, he has developed an interest in many different areas like functional programming or computer graphics, and some other less-known ones such as topological data analysis, computer algebra systems or type theory.
Probabilistic streaming algorithms: accuracy and speed at low cost
In recent years, there has been extensive study on algorithms able to process big amounts of streaming data with little memory footprint. Probabilistic algorithms (and probabilistic data structures) come with counter-intuitive good results: they are fast, light and have great accuracy. While the term “probabilistic” usually sounds like randomly picking or discarding data, the fact is that these algorithms show real cleverness and use randomness in unexpected ways to achieve very precise results that use a tiny part of the memory that would need an exact algorithm.
In this talk we will explore the role and functioning of probabilistic algorithms in some common scenarios for massive data like count-distinct or heavy hitters problems. With the help of some examples, we will discuss the desirable properties of a good streaming algorithm, before diving into HyperLogLog++ and Count-Min Sketch algorithms which, despite having simple core ideas, cope with major tasks - like analyzing all search queries on Google in real-time.
Dave Aronson
Software Development Consultant at Codosaurus
Dave Aronson
Software Development Consultant at Codosaurus
Dave is a semi-retired software development consultant (writing code and giving advice about it), with 37 years of professional experience in a wide variety of languages, systems, frameworks, techniques, domains, etc. He is the T. Rex of Codosaurus, LLC (his one-person consulting firm, which explains how he can get such a cool title, at https://www.Codosaur.us/) near Washington, DC, USA. His main focus in software is to spread the gospel of quality, including defining what that even means. In his spare time, he makes mead and teaches others how.
Tight Genes: Intro to Genetic Algorithms
Yes, that’s right, geneTic, not geneRic. Genetic algorithms are a way to “evolve” solutions to a problem, similar to real-world biological evolution. This often reveals great solutions that humans probably would never have thought of, such as the twisty NASA ST5 spacecraft antenna, developed by a genetic algorithm in 2006!
This talk will explain the concept and its terms, and then walk you through some examples, including creating a simple generic genetic-algorithm “runner”, and multiple algorithms for it to run, such as characters to fit D&D classes and mead recipes to yield specified levels of sweetness and strength.
Johnny Willer
Software Engineer at Ocado Technology
Johnny Willer
Software Engineer at Ocado Technology
Johnny Willer is a software engineer at Ocado Technology, currently working in the Payments domain within the company’s ecommerce stream. He has a Bachelor’s degree in Computer Science, is Java 11 OCP certified and strongly believes that quality software should be the standard rather than the exception. With ten years of experience in software craftsmanship, Johnny has worked on a variety of domains, both in the private and public government sectors. On a more personal note, he is a chess enthusiast and flower stick juggler.
Hexagonal Architecture and Monolith decomposition
Would you like to know how to decompose a monolith into a modular one using design practices like hexagonal architecture driven by use case semantics? This powerful technique enables an application codebase to evolve while minimising the risks of highly coupled, low cohesive modules and fragile tests.
Payments are part of everyday life. Whatever we buy, we need to pay for, right? For consumers, this seems simple, but only engineers understand the deep complexities inside payments – and consumers’ low tolerance for error. After launching the world’s first pure-play grocery retailing website 20 years ago, we’ve developed the Ocado Smart Platform (OSP), our end-to-end online grocery fulfilment solution, adopted by 12 of the world’s most forward-thinking retailers. This fast growth comes with the potential challenge of keeping the codebase clean, cohesive, and low coupled while following market-leading architectural principles like SOLID. To address this, we use a new, innovative architectural approach, based on Hexagonal Architecture driven by Use Case semantics. This architectural approach is being deployed gradually during a Monolith Decomposition. We use the Feature Flag technique quite extensively to select different infrastructure components and achieve other behaviours at runtime. In this session, you’ll also get some valuable tips and tricks to apply this style in your codebase.
Key takeaways:
- Use of the Feature Flag technique to achieve multiple behaviours at runtime
- Restructuring a core microservice from a Monolith into a Modular Monolith
- Application of Hexagonal Architecture to achieve multiple infrastructure adapters at runtime
- Understand the reason why Service Layer Based Development (SLBD) is not a feasible alternative
- Discover how UseCase classes improve on SLBD
Joaquín Guillén
Head of Architecture at The Workshop
Joaquín Guillén
Head of Architecture at The Workshop
Joaquin leads the Architecture team at The Workshop, where he helps design and build highly scalable distributed systems. With over 20 years of experience in software engineering across multiple industries, Joaquin has a particular passion for evolutionary architecture.
Continuous Evolution: Using Fitness Functions to Drive Platform Modernisation
In the era of continuous evolution, software products need to keep up with the rapid pace of change. The products we design and develop must evolve, adapting to support business growth and align with the latest technological changes. In this way, we ensure they can be maintained and built upon, instead of simply becoming legacy.
In this talk, Joaquin will explore how the principles of evolutionary architecture can help us visualise software product evolution, and how these visualisations can be used to prioritise and drive platform modernisation initiatives.
Chris Ford
Head of Technology at Thoughtworks
Chris Ford
Head of Technology at Thoughtworks
Chris is the Head of Technology for Thoughtworks Spain and an experienced architect and technical advisor. His career has taken him from Australia to the UK, India, Uganda and now Spain. As a consultant, he helps clients with architecture, agile development and organisational effectiveness. Chris was a technical reviewer for Zhamak Dehghani's 2022 book 'Data Mesh'. His personal interests include music-as-code, the art of using functional programming as musical notation.
Data Mesh 101
Data Mesh is a new socio-technical approach to data architecture, first described by Zhamak Dehghani and popularised through a guest blog post on Martin Fowler's site (https://martinfowler.com/articles/data-monolith-to-mesh.html). Since then, community interest has grown, due to Data Mesh's ability to explain and address the frustrations that many organisations are experiencing as they try to get value from their data. The 2022 publication of Zhamak's book on Data Mesh (https://www.oreilly.com/library/view/data-mesh/9781492092384/) further provoked conversation, as have the growing number of experience reports from companies that have put Data Mesh into practice. So what's all the fuss about? On the one hand, Data Mesh is a new approach in the field of Big Data.
On the other hand, Data Mesh is an application of the lessons we have learned from domain-driven design and microservices to a data context. This talk will explain how Data Mesh relates to current thinking in software architecture and the historical development of data architecture philosophies. They will outline what benefits Data Mesh brings, what trade-offs it comes with and when organisations should and should not consider adopting it.
Pablo Porto
Lead Developer at Thoughtworks
Pablo Porto
Lead Developer at Thoughtworks
Pablo is a Lead Developer for Thoughtworks Spain's Data and Artificial Intelligence Service Line. He is a skilled practitioner with a considerable breadth of experience, from infrastructure to microservices to data engineering. Pablo helps startups build MVPs, scale-ups evolve their teams and delivery practices and big enterprises build reliable infrastructure in the cloud. His current focus is to help his clients build robust, testable and maintainable data architectures.
Data Mesh 101
Data Mesh is a new socio-technical approach to data architecture, first described by Zhamak Dehghani and popularised through a guest blog post on Martin Fowler's site (https://martinfowler.com/articles/data-monolith-to-mesh.html). Since then, community interest has grown, due to Data Mesh's ability to explain and address the frustrations that many organisations are experiencing as they try to get value from their data. The 2022 publication of Zhamak's book on Data Mesh (https://www.oreilly.com/library/view/data-mesh/9781492092384/) further provoked conversation, as have the growing number of experience reports from companies that have put Data Mesh into practice. So what's all the fuss about? On the one hand, Data Mesh is a new approach in the field of Big Data.
On the other hand, Data Mesh is an application of the lessons we have learned from domain-driven design and microservices to a data context. This talk will explain how Data Mesh relates to current thinking in software architecture and the historical development of data architecture philosophies. They will outline what benefits Data Mesh brings, what trade-offs it comes with and when organisations should and should not consider adopting it.
Nikhil Barthwal
Senior Software Engineer at Facebook
Nikhil Barthwal
Senior Software Engineer at Facebook
Nikhil Barthwal is passionate about building distributed systems. He has several years of work experience in both big companies & smaller startups and also acts as a mentor to several startups. Outside of work, he speaks at international conferences on several topics related to Distributed systems & Programming Languages. You can know more about him via his homepage www.nikhilbarthwal.com.
Modeling and Verification of Concurrent & Distributed systems
The majority of Distributed systems are designed as untestable whiteboard drawings. This leads to design flaws that go unnoticed. Human intuition has its limits and these systems operate beyond those limits. This talk describes how we can use mathematical models and to find & eliminate these flaws.
Distributed & concurrent systems have grown exponentially in popularity in recent years. However, the vast majority of these systems are designed as untestable whiteboard drawings. This leads to fundamental design flows that go unnoticed in the design phase leading to hard to find bugs that are expensive to correct.
This talk is about TLA+, a high-level language used for modelling concurrent & distributed systems using simple mathematics. These models can be thought of as blueprint design of software except it is exhaustively-testable.
Human intuition has its limits and most of these systems operate at a scale which is well beyond what we humans can comprehend. TLA+ specifications are written in formal language and they are amenable to finite model checking. The model checker finds all possible system behaviors up to some number of execution steps and examines them for violations of desired invariance properties such as safety and liveness.
The objective of the talk is to demonstrate how TLA+ can be used to eliminate design flaws before system implementation is underway.
Hila Fish
Senior DevOps Engineer at Wix
Hila Fish
Senior DevOps Engineer at Wix
Hila Fish is a Senior DevOps Engineer at Wix, with 15 years of experience in the tech industry.
AWS Community Builder, and a public speaker who believes the DevOps culture is what drives a company to perform at its best and talks about that and other DevOps/Infrastructure topics at conferences.
She carries the vision to enhance and drive business success by taking care of its infrastructure.
In her spare time, Hila is a lead singer of a cover band, giving back to the community by co-organizing DevOps-related conferences (Inc. "DevOpsDays TLV" & "StatsCraft" monitoring-focused event), providing mentorship and managing programs in “Baot” (The largest technical women’s community in Israel), and enjoys sharing her passion and knowledge wherever she can, including across diverse technology communities, initiatives and social media.
Technical Documentation - How Can I Write Them Better and Why Should I Care?
Data collection done by people is a wasteful act and could result in duplicated work by different people. Gathering info for tasks, or for the ability to maintain code or infrastructure - Documentation plays a crucial part in that.
In this talk, I’ll show you a structured way to write a technical doc, without being a technical writer - So everyone could do it to their best ability. I’ll explain why you should care about these docs, and how eventually it serves your best interests (Yes, more than 1). If you want to save your time and other people’s time - Writing documentation well could have a great impact on that.
Sergey Bykov
SDE at Temporal Technologies
Sergey Bykov
SDE at Temporal Technologies
Sergey Bykov is responsible for the architecture of Temporal Cloud, a hosted service that is helping businesses, from large enterprises to tiny startups, to build invincible applications. Prior to joining Temporal Sergey was one of the founders of the Orleans project at Microsoft Research and led its development for over a decade. The mediocre state of developer tools for cloud services and distributed systems at the time inspired him to join the Orleans project in order to qualitatively improve developer productivity in that area. The same passion brought him to Temporal.
Inception or deja vu all over again
In this talk, Sergey will share his team’s experience of bringing Temporal Cloud to life. He will explain architecture of the service, with the primary focus on building its Control Plane. Sergey will cover approaches and patterns used in the process and lessons learned from successes and mistakes along the way. This should be of interest to engineers building or thinking of building multi-tenanted hosted services, which has become the most sustainable way of monetizing open source server products.
Gonzalo G. Jiménez Fuentes
Cloud Native Consultant at his own
Gonzalo G. Jiménez Fuentes
Cloud Native Consultant at his own
Gonzalo G. Jiménez Fuentes works as Consultant building platforms and tools to easily work in Cloud environments. He enjoys solving problems using software and he is very passionate about automation and system integration. His experience goes from small startups to big enterprises.
"All Matters Rust" Unconfakerence
Based on the core concepts of unconferences and open spaces we have created a new disruptive, scalable, fully decentralised, secure, and resilient session called unconfakerence.
While unconferences are supposed to be participant-driven meetings without an agenda or topic to be discussed, in the case of this unconfakerence there will be a predefined topic about Rust (aka RUST for some C++ programmers) on which different suggestions and subtopics will be raised.
The organisation will be led by Dr Roland Kuhn with the help of Arno Schots, and Gonzalo G. Jimenez, members of the JOTB organizing committee.
The Law of 2 feet prevails
If at any time during the session, a person feels that they are not learning or contributing, they have a responsibility to get up and move, use both feet to move on to something interesting.
Arno Schots
Cloud Solution Engineering Director at Oracle
Arno Schots
Cloud Solution Engineering Director at Oracle
Arno Schots is a Cloud Solution Engineering Director at Oracle, leading a team of Cloud Architects in EMEA. These teams are responsible for helping customers move to the cloud. Combining Enterprise Architecture background with hands-on engineering and development experience, he is passionate about sharing this knowledge with customers and technology audiences around the world. You can find him on Malt.es for free-lance engagements. In his free time, he can be found outdoor doing sports like trail running, cycling and surfing. Last but not least, he is co-organiser of the great Jonthebeach event.
"All Matters Rust" Unconfakerence
Based on the core concepts of unconferences and open spaces we have created a new disruptive, scalable, fully decentralised, secure, and resilient session called unconfakerence.
While unconferences are supposed to be participant-driven meetings without an agenda or topic to be discussed, in the case of this unconfakerence there will be a predefined topic about Rust (aka RUST for some C++ programmers) on which different suggestions and subtopics will be raised.
The organisation will be led by Dr Roland Kuhn with the help of Arno Schots, and Gonzalo G. Jimenez, members of the JOTB organizing committee.
The Law of 2 feet prevails
If at any time during the session, a person feels that they are not learning or contributing, they have a responsibility to get up and move, use both feet to move on to something interesting.
Scott McAllister
Developer Advocate at ngrok
Scott McAllister
Developer Advocate at ngrok
Scott McAllister is a Developer Advocate for ngrok. He has been building software in several industries for over a decade. Now he's helping others learn about a wide range of web technologies and incident management principles. When he's not coding, writing or speaking he enjoys long walks with his wife, skipping rocks with his kids, and is happy whenever Real Salt Lake, Seattle Sounders FC, Manchester City, St. Louis Cardinals, Seattle Mariners, Chicago Bulls, Seattle Storm, Seattle Seahawks, OL Reign FC, St. Louis Blues, Seattle Kraken, Barcelona, Fiorentina, Borussia Dortmund or Mainz 05 can manage a win.
As Malaga CF hardcore supporters, the organisation of JOTB condemns cheering in favour of Robussia Dortmund.
Building Ingress - From Concept to Connection
At first glance, ingress is an easy concept: you route traffic from the wider world into your cluster. As you layer on SSL and load balancing, the principles stay the same and everything works with minimal thought and effort. But as your infrastructure grows, your clusters grow, the interactions get more complex, and your security requirements explode. In this session, I’ll walk you through how we designed and built an Ingress Controller and have converted our clusters to use it in production to support millions of requests. It wasn’t easy but running it as an open source effort from the start encouraged our team and customers to review, explore, and consider situations outside our original plans.
Álvaro López
R&D Engineer / Innovation at Fortris
Álvaro López
R&D Engineer / Innovation at Fortris
Alvaro Lopez has +10 yrs of expertise as a software engineer in the video game industry. In recent years, he has turned his attention to decentralized technologies in terms of cybersecurity and scalability. Now, he holds the position of R&D Engineer at Fortris and is pursuing a Ph.D. at the University of Malaga in decentralized technology and its cybersecurity implications.
The price of scalability in blockchain
This talk examines the challenges of balancing security and performance in blockchain technology. It covers the costs of improving performance, known challenges in blockchain platforms, and the issue of centralization in decentralized systems.
Marcin Szymaniuk
CEO at Tantus Data
Marcin Szymaniuk
CEO at Tantus Data
Marcin is a CEO and Hands on Data Engineer at TantusData. He has a lot of hands-on experience with technical problems related to Big Data (Clusters with hundreds of nodes) as well as practical knowledge in business data analysis and Machin Learning. Companies Marcin has worked for or consulted for include: Spotify, Apple, Telia and small startups.
Go big or go…well not one too many. Aka applying machine learning in production
We’ll quickly define a ‘model on production’. There is a myriad of definitions people use right now. A variance wholly justified. Because the exact way we do define it depends. Among other factors on the size of the company, the number of models, the properties of the data, and so on. While we’re at it we’ll also answer some pinning questions such as: Is it ever OK to duct-tape the model deployment process? What about using shortcuts and opting for some manual work? Pithy answers. We’ll pause to take on any questions. If there will be too many to fit, we’ll provide contact information and make sure to address all enquiries after the presentation. Now practice makes perfect. Or nearer perfection at least. So we’ll briefly present two cases of solutions solving the same problem for two clients, both implemented with a vast difference. We’ll explain why. Both solutions required optimising search engine for results that translate into higher revenue. Yet, the companies are on opposite sides of the scale. One a huge, mature retailer, another a much younger and smaller online-booking business. The cost must always be justifiable. So, using these cases we’ll succinctly show how to fit a solution to match the context of the organisation. How to utilise it well. Next, quickly going through some details of the models and infrastructures, we’ll explain the reasoning behind the critical decisions as well as highlight the pros and cons of both approaches.
Chanuki Illushka Seresinhe
Head of Data Science at Zoopla
Chanuki Illushka Seresinhe
Head of Data Science at Zoopla
Chanuki Illushka Seresinhe is the Head of Data Science at Zoopla, where she manages data scientists both at Zoopla and Hometrack. She is also the founder of beautifulplaces.ai . Chanuki has a PhD in Data Science from the University of Warwick. Her research at the University of Warwick and the Alan Turing Institute, which involved understanding how the aesthetics of the environment influences human well-being, has been featured in the press worldwide including the Economist, Wired, The Times, BBC, Spiegel Online, Guardian, Telegraph, Scientific American, Newsweek and MIT Technology Review. Prior to pursuing a career in data science, she previously ran her digital design consultancy for over eight years.
Unravelling insights about places with computer vision
Images are a rich source of information about our environment, and deep learning has enabled us to extract insights from them with unprecedented accuracy and efficiency. In this talk, we will explore how deep learning can be used to gain insights about places, using images as the primary source of data.
Specifically, I will explain how I used AI to analyse crowd-sourced ratings of over 200,000 images of Great Britain from the online game Scenic-Or-Not to help us understand what beautiful places are composed of. I also trained an AI algorithm to be able to tell how beautiful a place is.
I will also talk about how we are using Deep Learning at Zoopla to gather interesting insights about properties. Ultimately, this talk will demonstrate how deep learning can revolutionise the way we analyse and understand places.
Olga Nosenko
Senior Software Engineer in Data Analytics and Visualization at Epam
Olga Nosenko
Senior Software Engineer in Data Analytics and Visualization at Epam
Olga Nosenko works 8-9 hours per day as a Senior Software Engineer working with Data Visualisation. Afterwards, she switches to a wine enthusiast. She is the owner of a golden retriever and a travel geek. She has a background in Supply Chain Management but once upon a time, she found herself ready to start all over and switch to IT. And she is happier than ever :)
Unleash your superpower: you do more than others think of you
Have you ever found yourself in a situation where your work was left unseen? Well, it’s easy to receive recognition for putting out fires, rather than preventing ones. I will tell you how I faced the need of making invisible work recognized during my assignment on a Data Visualisation project, give hints on how to do that and why it is crucial for you.
Nevertheless, we work in a fast-developing world and a huge number of new technologies and tools arise every month, what we actually seek is simple solutions to complex problems. But reality hits differently and is often left unseen as well as work that is done by employees while only the final part is presented to stakeholders and other team members.
I will provide you with some hands-on experience demonstrating full-cycle product development of Tableau Dashboard from Design proposition to Server administration.
I will cover examples of some non-trivial tasks of Visualisation engineers conducted in order to enhance the final product and automate the development.
Guillermo Del Río Párraga
DevOps Team Lead at Xebia Functional
Guillermo Del Río Párraga
DevOps Team Lead at Xebia Functional
Guillermo is a Telecom engineer with a backend development background, who transitioned to DevOps when someone had to deploy those new pesky microservices. He is also a Kubernetes administrator and is well-experienced in coaching developer teams to adopt good CI/CD practices.
GitOps CD, lessons learned
GitOps and CD are two terms that you can throw around in a meeting and colleagues will be impressed. "They really must know their stuff", people quietly think. It is widely recognised what improvements and general advantages in the process they bring to a DevOps flow.
In this talk we will explore some of the more advanced implementation quirks, using ArgoCD as an example:
- Release versioning and multi-environment.
- No more Helm hooks?
- Orchestrating deployments: sync waves and waves.
- Another layer of GitOps: automatic app creation.
- Sync windows.
Alasdair Brown
Developer Relations Lead at Tinybird
Alasdair Brown
Developer Relations Lead at Tinybird
Alasdair leads Developer Advocacy at Tinybird. He has spent the past 10 years in Cyber Security & Big Data, building some of the largest data platforms in the world, enabling financial institutions, critical national infrastructure and defense agencies to work in real time, at scale, to detect and analyze threats. He joined Tinybird in 2022 because “you shouldn’t need $2 billion and 400 engineers to do real time”.
Elena Torró
Software Engineer at Tinybird
Elena Torró
Software Engineer at Tinybird
Elena is a Software Engineer with experience in both Frontend and Backend development. In the last six years, she's been working on enabling developers to use technical data products, ranging from map libraries at CARTO to real-time analysis tools at Tinybird. In addition to her passion for technology, she practices Karate and Paddle Surf, creates riddles, and writes Science Fiction stories.
Iván De Prado Alonso
Head of Artificial Intelligence at Freepik Company SL
Iván De Prado Alonso
Head of Artificial Intelligence at Freepik Company SL
Iván de Prado Alonso is an expert in artificial intelligence and big data. He currently serves as the Head of Artificial Intelligence at Freepik Company, a technology company that produces and distributes graphic assets. He has applied AI to various fields, including agriculture, web scraping, and image retrieval. His curiosity has led him to get involved in startups, found a company, delve deep into big data and distributed systems, and even study economics. However, he has now found a passion for artificial intelligence.
AI Image Generation: From Text to Reality
In this talk, I will explore the fascinating world of AI image generation, a field of research that can produce realistic, creative, and diverse images from text, sketches, or other inputs. I will introduce the main concepts and techniques of AI image generation and showcase some of the best AI image generators. I will also share our own work and experience in developing AI image generators for Freepik, a leading platform for free graphic resources. I hope to inspire you with the amazing capabilities and possibilities of AI image generation, and invite you to join me on this exciting journey.
Antonio Mendoza Pérez
Developer Success Engineer at Temporal Technologies
Antonio Mendoza Pérez
Developer Success Engineer at Temporal Technologies
Antonio is a Developer Success Engineer at Temporal, where he helps big companies and startups to build invincible applications. Before joining Temporal, he worked for 10+ years as a software developer with Java, Typescript and Python across several industries.
He contributes to the Serverless Workflow Specification, where he maintains Typescript and Python SDKs.
Sammy Landstrom
Senior Director, CellRebel Data Engineering at Ookla
Sammy Landstrom
Senior Director, CellRebel Data Engineering at Ookla
Sammy has spent his career working in the borderland between business and tech helping organizations understand, structure and achieve value from data, more recently with a focus on connectivity insights.
Sammy has experience spanning from working deep within the technology side with cloud architecture and development to the business side including Agile Product management, Team management, Business Analysis, Design and Information Architecture.
His areas of expertise include: Cloud architecture, Snowflake, Amazon AWS, Clickhouse, Microsoft Azure, Kimball methodology and dimensional modeling, Agile, Scrum, Service Oriented Architecture, .Net, Python
Connectivity insights with geospatial analytics on Clickhouse
In this talk, Sammy Landstrom will tell the story of our journey evaluating several on the fly analytics engines, to selection and implementation of the Clickhouse engine. We will share our successes, and some challenges we’ve faced, and show some examples of what a hundred billion connectivity measurements per day looks like when measuring the state of global connectivity.
Vasja Çollaku
Senior Java Engineer at LeoVegas Group
Vasja Çollaku
Senior Java Engineer at LeoVegas Group
Vasja is from Albania, where she did her bachelor's degree in Computer engineering. In 2017 she got the opportunity to do her master's degree in Sweden, in the lovely town of Västerås, and after finishing that, she got to start at LeoVegas, which has been a love story ever since 2018. She started as a junior developer and worked her way up to senior, and she has been working extensively with Responsible Gaming since the beginning. I am a big fan of Inter Milan and love to write book reviews in my free time.
Responsible Gaming in iGaming
How can we ensure responsible Gaming is always present and accessible in our industry?
Ancizar Arenas
Engineer Manager at Saber.tech
Ancizar Arenas
Engineer Manager at Saber.tech
Ancizar is a seasoned Software Engineering Manager with over a decade of experience in the industry. He has a proven track record of building and delivering complex software projects across various industries and products such as travel & tourism, lottery and gaming, all whilst fostering a culture of innovation and collaboration within his teams.
As a leader, Ancizar believes that a team's success is directly tied to the happiness and growth of its members. He fosters an environment of trust, respect, and collaboration where his team members can thrive and develop their skills to the fullest.
His passion for servant leadership and technology has not only allowed Ancizar to make a positive impact in businesses but also in the teams that he has led and their team members, many of who have grown and developed into successful leaders themselves.
Our journey to distributed architecture - scaling from monolith to SCS
Our goal is to share our experiences, both from a technical and business perspective, with the community and inspire others who are facing similar challenges. Join us for an informative and inspiring talk on scaling software solutions and empowering businesses.
We'll take you on our journey of scaling to a distributed architecture, and how we tackled the challenges that came with it.
We'll cover how we scaled and organised development teams to match the evolving architecture and business needs, and the lessons we learned along the way.
We’ll also cover how we addressed these business needs from an architectural perspective by talking about how we adopted a self-contained system architecture whilst applying domain-driven design principles.
Javier Serrano Velázquez
Engineering Manager at Saber.tech
Javier Serrano Velázquez
Engineering Manager at Saber.tech
Javier is an experienced Software Engineering Manager with over 15-years’ experience in designing and developing innovative software solutions for businesses of all sizes.
He is passionate about leveraging technology to empower businesses to achieve their goals and drive growth.
As a servant leader, Javier is committed to putting the needs of his team and customers first. He believes in leading by example and empowering his teams to take ownership of their work while providing the support and resources needed to help them succeed. He is dedicated to building strong relationships with his stakeholders, understanding their unique needs and challenges, and working collaboratively to develop tailored software solutions that drive real business results.
Our journey to distributed architecture - scaling from monolith to SCS
Our goal is to share our experiences, both from a technical and business perspective, with the community and inspire others who are facing similar challenges. Join us for an informative and inspiring talk on scaling software solutions and empowering businesses.
We'll take you on our journey of scaling to a distributed architecture, and how we tackled the challenges that came with it.
We'll cover how we scaled and organised development teams to match the evolving architecture and business needs, and the lessons we learned along the way.
We’ll also cover how we addressed these business needs from an architectural perspective by talking about how we adopted a self-contained system architecture whilst applying domain-driven design principles.
Antonio Pinazo
Senior Financial Auditor at EY
Antonio Pinazo
Senior Financial Auditor at EY
Antonio is a Senior Financial Auditor at EY who is dedicated to achieving maximum efficiency in his work.
By implementing a digital audit platform, data analytics, and automation tools, he has been instrumental in expanding his team from five to over fifty members in just three years.
In pursuit of further growth, Antonio is now leading efforts to integrate technologies such as artificial intelligence into the audit process.
Thanks to his focus on efficiency, Antonio is able to make time for a variety of hobbies, including American Football and dancing.
Fraud and Error Detection: Applications of Artificial Intelligence in a Traditional Sector like Financial Auditing.
AI has had a significant impact on financial auditing, revolutionizing the way audits are conducted and improving their effectiveness. AI has enabled auditors to process and analyze large amounts of financial data more quickly and accurately than ever before. By using AI-powered tools such as machine learning algorithms and natural language processing, auditors can detect anomalies and potential fraud that may have gone undetected using traditional auditing methods.
Taras Pedchenko
Engineering & Senior Delivery Manager at Intellias
Taras Pedchenko
Engineering & Senior Delivery Manager at Intellias
Taras started his IT journey in 2004, having been holding leadership roles since 2014. His professional experience comprises both working with large-scale international corporations and fast-growing start-ups with a particular focus on helping businesses turn their ideas into highly valuable digital products. At a personal level Taras is driven by the values of people-centricity, excellence and innovation.
Cloud cost optimization: money is not a problem, but the amount is
During his talk, Taras will guide the audience through real-life cases and best practices on how to resolve issues related to the high cost of cloud services. In particular, he will focus on state-of-the-art approaches to designing and operating cloud environment cost-efficiently, as well as guidelines on how to build transparency and understanding among engineering teams and company stakeholders when addressing a sensitive topic of cost-optimization.
Guillermo Ruiz
Senior Developer Advocate at AWS
Guillermo Ruiz
Senior Developer Advocate at AWS
Guillermo is a Senior Developer Advocate, leading AWS developer awareness and adoption efforts in Iberia, engaging with builders, AWS Community and third-party whitespace technical audiences. In this role, G focuses on creating and delivering compelling programs and content to inspire builders on what’s possible with AWS, acting as tech strategist, and building long-term relationships with builders and communities.
Guillermo is an entrepreneur at heart and a builder. He is a former startup co-founder. Some of his crazy work includes designing real-time business analytics and AIOps services, Augmented Reality in Datacenter operations and Gaming sentiment analysis.