We are back with another episode of True ML Talks. In this, we dive deep into ML Platform and we are speaking with Tushar
Tushar is a seasoned MLOps leader with 20+ years of experience at top tech companies and a wide range of skills in product, business, engineering, and investment banking. He is also the co-founder of the worldwide IIT, artificial intelligence, and machine learning forum, and runs a very active Slack community for that as well.
IIT AI/ML Forum was started with the vision of creating a community where IITians working in AI/ML could share their knowledge, collaborate, and help each other. They believed that by working together, IITians could surpass any other engineering institute in the world.
The forum has been a huge success, with over 1800 members from all over the world. The forum has organized events, supported other organizations, and grown into a thriving community in its own right.
Tushar is particularly proud of three things that the forum has accomplished:
The combination of the Cloud computing, Transformers, Pre-training will be a major driver of innovation in AI and MLOps in the coming years. Particularly the potential of multimodal AI, which combines natural language processing and computer vision to solve complex problems.
Cloud computing has made AI more accessible and affordable for everyone. This has led to a surge of innovation in the field, as startups and individuals are now able to develop and deploy AI applications without having to invest in expensive infrastructure.
Transformers have revolutionized natural language processing and computer vision. Transformers are a type of neural network architecture that is able to learn long-range dependencies in data. This makes them well-suited for tasks such as machine translation and image recognition.
Pre-training is a technique where a large language model is trained on a massive dataset of text and code. This pre-trained model can then be fine-tuned for specific tasks, such as translation or question answering. Pre-training has significantly improved the performance of AI models on a wide range of tasks.
ChatGPT and generative AI have the potential to revolutionize many industries. He is particularly interested in the potential of these technologies to improve customer service, reduce fraud, personalize products and services, and improve healthcare.
Examples of specific applications of ChatGPT and generative AI in different industries:
LLMs are still in their early stages of development, but they have the potential to revolutionize risk assessment in the financial services industry:
LLMs can process more data, faster. Risk assessment models traditionally rely on a limited amount of data, such as credit scores and income. LLMs can process much more data, such as spending patterns, buying behavior, and online behavior. This allows them to create more accurate risk assessments.
LLMs can consider associative factors. In addition to individual factors, such as credit score, LLMs can also consider associative factors, such as the company a person works for and the industry they work in. This can help them to create more comprehensive risk assessments.
He believes that there will be three types of players in the ecosystem:
📌
Electical Power Industry:In the electrical power industry, there are generators, transmission lines, and distributors. In the LLM industry, Tushar sees foundation model builders as generators, cloud computing providers as transmission lines, and startup companies as distributors.
There will be a space for both closed and open source LLMs. Closed source models will be preferred by large enterprises that need production-ready solutions with support. Open source models will be preferred by smaller companies and researchers who need more flexibility and customization.
There will be a need for middleware to help developers use LLMs more easily and efficiently. Middleware can provide features such as model management, fine-tuning, and monitoring.
It is imperative to view LLMs as tools that can either amplify human capabilities or pose risks, depending on their application. Like any tool, the use of LLMs is shaped by human choices and intentions. They hold the potential to Advance medical treatments, Foster innovative educational programs, Automate tasks currently performed by humans. However, they can also Generate deepfakes, Spread misinformation, Manipulate individuals.
Even as LLMs grow in sophistication, they will always fall short of fully grasping the nuances of human values. Consequently, humans retain a pivotal role in ensuring that LLMs align with our values. This includes, Establishing ethical guidelines for LLM development and usage, Educating the public about LLM benefits and risks, and Recognizing that humans possess the unique capacity to think creatively and find innovative solutions, while LLMs are constrained by their training data.
When it comes to constructing generic RAG systems, AWS and startups each bring their own distinct advantages and challenges to the table.
AWS Strengths: AWS is well-placed to develop generic RAG systems due to its substantial customer base and a wide range of services that can support RAG. For example, AWS offers SageMaker, a machine learning platform for training and deploying RAG models. Additionally, AWS provides various data storage and processing services ideal for RAG workflows.
AWS Weaknesses: AWS might not match the agility of startups in terms of swiftly developing and launching new products. Furthermore, AWS's focus may not be as specific as startups, especially in use cases like RAG for healthcare.
Startup Advantages: Startups excel in agility, allowing them to focus on specific use cases and rapidly innovate in the RAG domain. Their niche focus can lead to unique RAG solutions and innovations often overlooked by larger entities.
Startup Challenges: Startups often grapple with resource constraints, lacking the extensive customer base and service portfolio of AWS. Competing with AWS on price can be daunting due to the scale and resources of the tech giant.
Keep watching the TrueML youtube series and reading the TrueML blog series.
TrueFoundry is a ML Deployment PaaS over Kubernetes to speed up developer workflows while allowing them full flexibility in testing and deploying models while ensuring full security and control for the Infra team. Through our platform, we enable Machine learning Teams to deploy and monitor models in 15 minutes with 100% reliability, scalability, and the ability to roll back in seconds - allowing them to save cost and release Models to production faster, enabling real business value realisation.
Join AI/ML leaders for the latest on product, community, and GenAI developments