Series of talks by Dojo employees and Google Startup Program manager on how to include diversity in product development and career tips for miniorities in tech.
Series of talks and workshop on Generative AI with Gemini at the center, third party cookies and flutter for mobile development.
Workshop on Agent Builder a no code LLM framework from google to orchestrate Agent workflows of LLM.
Series of talks on Generative AI concepts, with a talk from Nishi Ajmera (Lead Engineer @ Publicis) on Enhancing similarity search systems using RAG.
A hackthon organized by NHS to present a startup Idea in the medical domain. Here is a picture of my team and I brainstorming on how on our Radiologist Model as Service startup to be sold to clinics. Our team consisted of a couple who flow in from India to write a medical entrance exam and my friend Mark. We did not win though.
First talk from Dr. Jodie Burchell on Measuring intelligence in AI models, with a review on a paper on sparks on AGI with GPTY4. Second talk from Chris Samiullah on developing AI applications using open source models like llama. The last talk was on model serving strategies from data collection, cleaning, model development and deployment with various tools.
Main talk from Victor Naroditskiy on improving generated text from LLms using RAG, know as Retrieval Augmented Generation and reduce hallucination of LLMs. Lighting talk was from Casper da Costa-Luis try us convince us Windows was better than Linux, I was like sorry bro that's for you. From the image you can tell many people did not show up due to the end of year festival activties.
Series of talks on career development, recruitement and growth within the tech industry. I met a lot of recruiters and made some new friends.
Adam MacVeigth from News Uk gave a talk about how the develop an Audio pipeline for text transcription using STT (Speech to Text) and TTS (Text to Speech) models using whisper on their Google Cloud infrastructure. One of the major challenges they faced was to splitting speakers and it was interesting to me because I had recently develop a similar POC application earlier for my podcast to transcribe audio to Text and make a summary, so I could relate. I won a Google Hat during the raffle draw and I presently preparing for the Google Data Engineer certification.
A talk from Marine Goseline of Taipy (open source Front-end+ Back for data applications) on how to turn Data/AI applications into full web apps using Taipy. The second talk was from Michael Natusch on 10 rules to mess your ML implimentation. The take away lesson from the first talk was ML and Data engineers find it difficult to develop client facing applications with Javascript frameworks and how tools like Streamlit and now Taipy hope to resolve this problem. I found Taipy intriguing but also realize the concept of experiment tracking was lacking at this time which is critical given the experimental nature of Data Science. The rules presented by Michael were intriguing as it seem obvious but somehow not always implemented and it also gave me good understanding of when and how to approach ML problems in cross functional team.
First talk was on evaluating distributional forecast using methods the Continuous Rank Probability Score (CRPS) and the Log Score and use them to pick the best forecasting algorithm using cross-validation. John Sandall one Pydata Organizers gave a talk about how to building your own Private ChatGPT using Streamlit, LangChain & Vicuna.
My cofounder and I were pitching our startup SKincare Pal to some UCL students to join us for an internship ion different technical roles as we we're seeking to start the development and needed some engineers on board. I was intrigued by some conversations with some students in the Masters program in AI and they we're eager to join our team. On the image I was explaining the technical details to some of the software engineering majors and the challenges we are solving at Skincare Pal with novel approach to skin care product recommendation using images.
I volunteered at PyData's London Flagship event, which is a 3 days conference on best practices in developing and shipping Data and Machine learning applications. It was rich day with speakers from startups and big tech corporations either selling a new service, demo, or trying to recruit. My main role was to register participants, chair different workshop sessions by presenting speakers and managing Q&A sessions after each talk. So much to learn and equally connected with some wonderful people in the ML community in London and from different places in the world.
This was a one-day event with different workshops and talks all going on in parallel. I chose to attend a workshop on recommendation systems using Multi-Armed Bandits because of my interest in recommendation systems, which turned out to be fascinating as I came away with an understanding of how Multi-Armed Bandits work using exploration and exploitation, and how to optimize the reward function. I also enjoyed the panel discussion with PwC on Responsible AI, though the panel did not have a clear answer to my question about making open source responsible without stifling innovation and regulating open source, which, in my humble opinion, contributes to innovation.
The main talk was by Sefik Serengil on doing Facial at billion Scale of data, in which he used deep face python library and the approximate nearest neighbour algorithm. It was interesting to use Faiss(Facebook AI Similarity Search) was used to search for an identity in seconds in a database with billions of data. This got quite intrigued about computer vision and it's possibilities in real world applications. The second talk was by Pavel Katunin and Anton Nikolaev on building scalable robots ensemble to collect big microscope imaging data
This workshop, organized by Snowflake with sponsorship from EDF, lasted one day and provided a remarkable opportunity to learn about the development of data pipelines and machine learning (ML) applications using Snowflake's platform. Over the course of three hours, I was able to create a small pipeline that ingested financial data into Snowflake's data warehouse, utilizing Snowpark—a framework that enables data processing and pipeline execution in languages such as Python, Java, and Scala. Additionally, we developed a regression model and built a client-facing application using Streamlit, which facilitated interaction with our ML outputs. During the workshop, I also had the pleasure of meeting an acquaintance in person—visible in the first picture—who had previously assisted me in troubleshooting a Docker issue I encountered while reaching out to him on LinkedIn
This meetup was quite interesting as we learned about from Alex Glaser of UK Health Security Agency how Data Science was used during the COVID lockdown period for infectious disease modeling to help in the development of response plans using Time Series Modeling and detecting symptons. The second talk was from Peter Vidos who presented ipyvizzu an open source framework for building animated data stories in Jupyter to present & sharing findings.