mastodon.ie is one of the many independent Mastodon servers you can use to participate in the fediverse.
Irish Mastodon - run from Ireland, we welcome all who respect the community rules and members.

Administered by:

Server stats:

1.6K
active users

#datascientist

2 posts2 participants0 posts today

Want to truly understand how AI learns to make better decisions?
:blobcoffee: Start with the concept of exploration vs. exploitation and dive into one of the most basic, but super important, ideas in Reinforcement Learning: Multi-Armed Bandits.
:blobcoffee: It's a simpler start to understand the basics behind. And it might change how you see AI.
towardsdatascience.com/simple-

Towards Data Science · Simple Guide to Multi-Armed Bandits: A Key Concept Before Reinforcement Learning | Towards Data ScienceHow AI learns to make better decisions and why you should care about exploration vs. exploitation

Why normalize databases?
Yesterday, my tutoring student asked me why databases need to be normalized at all. She said: “Wouldn’t it be easier to just have one big table with all the information?”

It’s a common first question when learning about relational databases.
At first, one big table (e.g. customer name, order date, product name, price) seems easiest.

I told her:
:blobcoffee: Because that quickly leads to data redundancy, anomalies, and integrity issues when inserting, updating, or deleting records.
:blobcoffee: Normalization means structuring data into separate, related tables, so that each fact is stored only once. This reduces redundancy & preserves consistency.

Can you remember learning to walk as a baby? You didn’t read a manual. Neither does an AI agent.

Reinforcement Learning (RL) isn’t about knowing the correct answer.
It’s about learning through trial and error, by interacting with an environment & receiving feedback.

That’s how AlphaGo defeated a world champion:
It first learned from expert games. Then it played against itself, millions of times, using RL to get better with each game. That’s how it mastered Go.

Want to really understand how RAG, vector search & chunking work?

Then stop reading theory and build your own chatbot.

This guide shows you how to create a local PDF chatbot using:

☕ LangChain

☕ FAISS (vector DB)

☕ Mistral via Ollama

☕ Python & Streamlit

Step-by-step, from environment setup to deployment. Ideal for learning how Retrieval-Augmented Generation works in practice.

👉 medium.com/data-science-collec

Comment “WANT” if you need the friends link to the article, as you don’t have paid Medium.

Data Science Collective · RAG in Action: Build your Own Local PDF Chatbot as a BeginnerBy Sarah Lea

🍕 Imagine trying two pizzerias and always going back to the one that seemed better. Sounds simple? Maybe.

But what if there’s another one in the city that’s even better – and you never tried it?

That’s what the greedy strategy does. It sticks to what worked best so far.

In Multi-Armed Bandits, a classic problem that helps us understand Reinforcement Learning, there are three common strategies to deal with this dilemma:
:blobcoffee: Greedy
:blobcoffee: ε-Greedy
:blobcoffee: Optimistic Initial Values

towardsdatascience.com/simple-

Towards Data Science · Simple Guide to Multi-Armed Bandits: A Key Concept Before Reinforcement Learning | Towards Data ScienceHow AI learns to make better decisions and why you should care about exploration vs. exploitation

Do you always go to the same café? Or do you try something new?

That’s the exploration vs. exploitation dilemma: Decision under uncertainty.

Multi-armed bandits model exactly that.

And this dilemma shows up everywhere: Recommender systems, A/B tests, online ads, even in human psychology.

Nobel Prize winner Daniel Kahneman called this one of the most fundamental cognitive patterns.

🎰 I explain what it is, why it matters, and how AI systems handle it.

:blobcoffee: Full article here: towardsdatascience.com/simple-

Towards Data Science · Simple Guide to Multi-Armed Bandits: A Key Concept Before Reinforcement Learning | Towards Data ScienceHow AI learns to make better decisions and why you should care about exploration vs. exploitation

I asked a Harvard postdoc a bunch of questions. These are the key takeaways:
:blobcoffee: The fundamentals still matters: math, statistics, coding & logic.
:blobcoffee: We need critical thinking & abstract reasoning.
:blobcoffee: Combining two disciplines leads to exciting projects and connections we couldn’t even imagine before.
:blobcoffee: A stay abroud is often an experience that changes more than our CV - they leave a mark in our mindset.
:blobcoffee: Two days in a postdoc's life rarely look exactly the same: Writing papers, training AI models, giving talks at conferences or running project meetings.
:blobcoffee: Success in research is more like a marathon with long phases without immediate feedback or visible success

-> Grab the full article: medium.com/ai-advances/ai-canc

Write in the comments "friend-link" if you don't have the paid Medium version.

#career #careergrowth #careeradvice
#scientists #research #researcher #AI #ki #computerscience #KI #datascience #datascientist #python #harvard #student

AI Advances · AI, Cancer Cells & a Master’s Student’s Journey to a Harvard PostdocBy Sarah Lea

Reinforcement Learning doesn’t tell you what’s right.
It only tells you how good your choice was.
No feedback on what to do. Only on how it went.

:blobcoffee: Example: A multi-armed bandit (like a slot machine with several levers). You don't know which lever is the best - you can only find out by trying it out. Exploring means giving up a known reward (from exploitation) — in hopes of finding a better one.

This balance between exploration and exploitation is the central dilemma in reinforcement learning.

:blobcoffee: A simple strategy is ε-greedy:
→ In 90% of cases you take the best known action
→ In 10% of cases, you try a different one by chance

In simulations, ε-greedy methods perform better in the long term than pure greed (always take the supposedly best) - because they master the “explore-exploit trade-off”.

What does a baby learning to walk have in common with AlphaGo’s Move 37?

Both learn by doing — not by being told.

That’s the essence of Reinforcement Learning.

In my latest article, I explain Q-learning with a bit Python and the world’s simplest game: Tic Tac Toe.

-> No neural nets.
-> Just some simple states, actions, rewards.

The result? A learning agent in under 100 lines of code.

Perfect if you are curious about how RL really works, before diving into more complex projects.

Concepts covered:
:blobcoffee: ε-greedy policy
:blobcoffee: Reward shaping
:blobcoffee: Value estimation
:blobcoffee: Exploration vs. exploitation

Read the full article on Towards Data Science → towardsdatascience.com/reinfor

Towards Data Science · Reinforcement Learning Made Simple: Build a Q-Learning Agent in Python | Towards Data ScienceInspired by AlphaGo’s Move 37 — learn how agents explore, exploit, and win

What does your Spotify account say about your music taste?
I was curious – so I built a little app to find out. Using Python, Streamlit and Plotly, I visualised my top tracks, favourite genres and the release years of the songs I listen to most.

What you need:
:blobcoffee: Access to the Spotify Web API
:blobcoffee: A bit of Pandas for analysis
:blobcoffee: A few lines of code with Plotly
:blobcoffee: Streamlit to build and run the app

→ The result? A personal, interactive music year in review 🎶

I put together a step-by-step guide (beginner-friendly) including code, screenshots, and how to deploy the app on Streamlit Cloud.

👉 Check out the full article here: bit.ly/3SF3VTw

#python #programming #technology #streamlit #spotify #Datavisualization #datascience #datascientist #opensource #api

Friends-Link: medium.com/data-science-collec

Data Science Collective · Visualize Your Spotify Data with Python and StreamlitBy Sarah Lea

You see a math formula you don’t immediately understand.

Your instinct? Stop reading.

Don’t.

That’s exactly what I told myself when I started reading Probabilistic Machine Learning – An Introduction by Kevin P. Murphy.

And it was absolutely worth it.

In my last article on Towards Data Science, I took a probabilistic view of supervised, unsupervised & reinforcement learning 👉 towardsdatascience.com/beyond-

Towards Data Science · Beyond Glorified Curve Fitting: Exploring the Probabilistic Foundations of Machine Learning | Towards Data ScienceAn introduction to probabilistic thinking — and why it’s the foundation for robust and explainable AI systems.

Imagine: Text corporas, curated by #libraries, whether #OpenAccess or licensed, ready to use out of the #IDE of a #DataScientist in a secure environment. No need to download lots of data, no need to worry about #AIBots. Here it is:
blog.oceanprotocol.com/free-co

We already delivered a Proof-Of-Concept for @CrossAsia and will continue...

Ocean Protocol · Free Compute-to-Data with Ocean Nodes VS Code ExtensionBy Ocean Protocol Team

Day 4 of the 5-day Generative AI intensive course by Kaggle and Google covered, among other things, Search Grounding:

:blobcoffee: With this technique, you can enrich generative AI models with up-to-date and reliable information from the internet — in this case by connecting them to Google Search.

That means the model no longer relies solely on its training data but also integrates information retrieved from search results.

:blobcoffee: So what's the difference to RAG (Retrieval-Augmented Generation)?

There’s no need to build and manage your own retrieval system (like a vector database), since the additional data is fetched via an API from a search engine.

Here is the link to the corresponding Kaggle notebook: kaggle.com/code/markishere/day

www.kaggle.comDay 4 - Google Search groundingExplore and run machine learning code with Kaggle Notebooks | Using data from No attached data sources
#ai#GenerativeAI#ki

:blobcoffee: What are embeddings?

Embeddings are numerical representations of things like words, sentences, images – or even customer profiles. They help models understand meaning and similarity between those objects.

:blobcoffee: Example:
The word "king" might be turned into a vector like [0.8, 0.1, …],
while "queen" becomes [0.81, 0.12, …]
→ Because these vectors are close, the words are semantically similar.

Embeddings are the backbone of:
📌 Similarity searches
📌 Recommendation engines
📌 Semantic search
📌 or RAGs (Retrieval-Augmented Generation)

And where are all these vectors stored? In vector databases, which let you search through embeddings lightning-fast.

:blobcoffee: This was the focus of Day 2 of the 5-Day Kaggle Generative AI Challenge: Embeddings & Vector Stores.