The #LLM kiss of death. #NaNoWriMo is closing down. Apparently not an April fools' prank.
https://www.reddit.com/r/nanowrimo/comments/1jog2tk/comment/mkripel/
The #LLM kiss of death. #NaNoWriMo is closing down. Apparently not an April fools' prank.
https://www.reddit.com/r/nanowrimo/comments/1jog2tk/comment/mkripel/
Call for participation: *ClimateCheck* Shared Task ( https://sdproc.org/2025/climatecheck.html)
@NFDI4DS members Raia Abu Ahmad, Aida Usmanova and Georg Rehm are organizing a shared task “Scientific Fact-Checking of Social Media Claims on Climate Change (ClimateCheck)” on July 31 or August 1st, 2025 in Vienna, Austria, hosted as part of the SDP 2025 Workshop.
Deadline for system submissions: May 16, 2025
#NLP
#LLM
#climatecheck
#socialmediaposts
#climatechange
#misinformation
#SDP2025
#ACL2025
#Vienna
#NFDI4DS
GitHub for Beginners: How to get LLMs to do what you want.
OMG, ChatGPT 5 is here and Gary Marcus is here for it!
https://open.substack.com/pub/garymarcus/p/breaking-gpt-5-news
Added a thing to generate code through Ollama if you set a flag on the command line.
It's... well, a work in progress.
I wonder if..
..tagging online, and for that matter offline content to indicate support for '#woke' ideals, you know, evil things like equality, anti-fascism, diversity, abortion access and especially #trans rights..
..would help in various ways, other than just the declaration of support. Such as:
- self censorship by bigots and their lackies depriving them of access to the value we create
- reduced scraping and appropriation (eg for #LLM)
- making it harder to spot and censor #wokeness
etc
AI doesn’t learn like we do. GoPro'd toddlers see mostly knees—but they can ID a dog after a few glimpses, while LLMs need billions of examples. Neuroscience suggests AI intelligence is gathered very differently than ours https://www.linkedin.com/posts/jonippolito_neuroscience-ailiteracy-generativeai-activity-7312811017170808832-dwUW
#GenerativeAI’s “black box” nature has brought quality assurance into sharp focus. From evals and benchmarking to guardrails, these approaches all play a crucial role in improving the reliability and accuracy of this technology.
Here are our top picks for understanding #LLM evals -
Large language model evaluation: A key to GenAI success: https://ter.li/cn7pbk
LLM benchmarks, evals and tests: A mental model : https://ter.li/juy34e
AI testing, benchmarks and evals: https://ter.li/smjohc
We had a wonderful time at the Leeds Data Science Meetup last week! A huge thank you to our amazing speakers, Abhinav Jindal and Andy Burgin, for sharing their insightful talks.
#Rstats #Networking #DataScience #Meetup #LLM
https://www.meetup.com/leeds-data-science-meetup/
Search Engine Journal: AI Researchers Warn: Hallucinations Persist In Leading AI Models. “Despite billions in research investment, AI factuality remains largely unsolved. According to the report, even the most advanced models from OpenAI and Anthropic ‘correctly answered less than half of the questions’ on new benchmarks like SimpleQA, a collection of straightforward questions.”
No, #AI frontier models don't "just guess words", it's far more complicated than that.
#Anthropic built an #LLM "brain scanner" (so far AIs have been black boxes).
According to Anthropic, "it currently takes a few hours of human effort to understand the circuits we see, even on prompts with only tens of words." And the research doesn't explain how the structures inside LLMs are formed in the first place.
This is a nice write up on my stance with #AI as well: https://sgnt.ai/p/hell-out-of-llms/. I have nothing against #machinelearning but against #llm's that are general can't be great at everything. It's what is the problem with #IT in general: everyone must be great at everything (#devops) and that will not be the case; #software #craftsmanship is suffering.
Efficient use of a #searchEngine 10 years ago:
Coming up with a phrase that would likely contained in an article that explains the thing I want to read
Efficient use of a search engine now:
Coming up with a phrase that avoids anything triggering #LLM #AI based content
@BenAveling @petealexharris @davidgerard
Next most probable word pattern matching isn't a brain; it's not artificially intelligent. It's more like a game of Mad Libs, and about as useful.
You know, we invented systems before there were computers.
'Forms' were on paper, rather than on screens.
An 'in tray' was an actual metal wire, or wooden tray, for paper letters, notes, memos and forms.
A database was called a 'filing cabinet'.
An 'interface' was a mail box.
A 'front end' was a person, with a job title like administrator, or clerk.
These systems were described, in excruciating detail, in procedure manuals.
The processes were run not by CPUs, but by people.
'Bugs' were when people made mistakes.
Systems were difficult to understand, even harder to diagnose, and very very hard to fix or change.
To change the way a department worked, for e.g. accounts receivable was so hard that most companies never even tried.
And yet somehow people are under the impression that it is the code that is the difficult bit about modern business systems.
So they try and make the code part easier.
#LowCode #LoCode #NoCode #AI #GenAI #LLM
It was never the code. Code was never the bottleneck.
I wonder if chains-of-thought I'm getting from these so-called reasoning AI's when I give them challenging combinatorics puzzles are exhibiting the characteristic #overthinking described in this recent publication
https://spectrum.ieee.org/reasoning-in-ai
#AI #reasoning #LLM #generativeAI
"If I’m 4 years old and my partner is 3x my age – how old is my partner when I’m 20?"
Do you know the answer?
An older Llama model (by Meta) said 23.
A newer Llama model said 28 – correct.
So what made the difference?
Today I kicked off the 5-day Kaggle Generative AI Challenge.
Day 1: Fundamentals of LLMs, prompt engineering & more.
Three highlights from the session: Chain-of-Thought Prompting
→ Models that "think" step by step tend to produce more accurate answers. Sounds simple – but just look at the screenshots...
Parameters like temperature and top_p
→ Try this on together.ai: Prompt a model with “Suggest 5 colors” – once with temperature 0 and once with 2.
Notice the difference?
Zero-shot, One-shot, Few-shot prompting
→ The more examples you provide, the better the model understands what you want.
HUH. Who woulda thunk that "ai" (LLMs) were going to be primarily a tool of oppression and wealth funneling, rather than liberation and help????
Gemini 2.5 Pro (experimental) still gets this question wrong. Its chain-of-thought seemed obsessed with counting white and black squares to test feasibility
Techbro proposals are ALWAYS SCAMS. No exceptions.
- NFTs
- Crypto
- ANYTHING blockchain
- LLMs
ALWAYS. A. SCAM.