So in @sirajraval's livestream yesterday he mentioned his 'recent neural qubit paper'. I've found that huge chunks of it are plagiarised from a paper by Nathan Killoran, Seth Lloyd, and co-authors. E.g., in the attached images, red is Siraj, green is original
Got a large dataset which takes many hours of training.
A method I used to quickly get a prototype, is sampling the data and creating a model that over-fits.
Then, once ready, i provide the rest of the data.
Saves time and money.
Thoughts?
*take caution with domain
Why would you want to use dimensionality reduction techniques?
Dimensionality reduction can allow you to:
• Speed up training by reducing the number of features
• Reduce memory usage
• Identify latent features that impact multiple features
Who's tired of hyper parameter tuning?
If you aren't, you're lying!
Search methods are the best way to find params without putting in too much of the effort that comes with hyper parameter tuning.
RandomSearchCV, GridSearchCV, GeneticSearchCV are some which can help you. :)
Not all tasks need neural networks, its surprising how many current business cases can be solved with other ML techniques as well as GROUPBY statements.
Unfortunately data sets with vague column names exist and dealing with them can get frustrating without proper documentation.
Not anymore.
By passing a Dict obj to the DataFrame.rename() function, you can tailor the data set as you like and make analysis fun. :)
Something different. People have DM'd asking for some general python tips too.
There is a misconception that to be good at coding you have to know all functions.
IMO, you should know how to apply functions no matter the language, use dir() or help() to look up functions.
Ever wanted to quickly analyze data you saw online or somewhere not as accessible.
pandas offers the read_clipboard() functions which allows you to make a data frame with ease.
From the screenshot you'll see I had data in Excel which I copied (Ctrl + C). :)
Following on from the previous memory management post, Chunks.
By passing the argument chunksize we can split our input data into chunks to avoid loading a large file into memory.
Note the change in the type() of both our data frames and the shape of our data frames. Enjoy :)
477 Followers 325 FollowingIndependent Alignment Researcher contracting with Anthropic on scalable oversight and adversarial robustness. I also work part-time at Speechmatics.
1.4M Followers 1K FollowingBuilding @EurekaLabsAI. Previously Director of AI @ Tesla, founding team @ OpenAI, CS231n/PhD @ Stanford. I like to train large deep neural nets.
956K Followers 765 FollowingProfessor at NYU. Chief AI Scientist at Meta.
Researcher in AI, Machine Learning, Robotics, etc.
ACM Turing Award Laureate.
1.2M Followers 279 FollowingWe’re a team of scientists, engineers, ethicists and more, committed to solving intelligence, to advance science and benefit humanity.
454K Followers 77 FollowingTensors and neural networks in Python with strong hardware acceleration. PyTorch is an open source project at the Linux Foundation. #PyTorchFoundation
386K Followers 624 FollowingLove Linux/Unix, open source, and programming? Into Sysadmin & DevOps? Follow us! Boost your IT career with daily new tools, apps, and humor ⤵️
1.3M Followers 1K FollowingCo-Founder of Coursera; Stanford CS adjunct faculty. Former head of Baidu AI Group/Google Brain. #ai #machinelearning, #deeplearning #MOOCs
717K Followers 288 FollowingTogether with the AI community, we are pushing the boundaries of what’s possible through open science to create a more connected world.
64K Followers 2K FollowingResearch Scientist at Google DeepMind (WaveNet, Imagen, Veo). I tweet about deep learning (research + software), music, generative models (personal account).
108 Followers 242 Following@ELLISforEurope PhD Student @DKFZ, @UniHeidelberg and @Cambridge_Uni. Working on dataset curation and limitations in medical imaging.
15K Followers 884 FollowingComet provides an end-to-end model evaluation platform for AI developers, with best in class LLM evaluations, experiment tracking, and production monitoring
25K Followers 1K FollowingFind me @[email protected] Professor at @OxCSML, @oxfordstats and Research Director at @GoogleDeepMind. All opinions are my own.
7K Followers 3K Following🛠️ Founder @AbideAI 👐 ML Engineer 👩💻☕
📚 Book Author: LLMOps (2025), ✍️ GPU Engg for AI Systems (2026)
💬🐦 Talk to me about LLMs, MLSys & GPU Training
2K Followers 2K FollowingWhat's the Kolmogorov Complexity / Minimum Description Length of a Reasoning Language Model? LLMs, AI/ML, Data Science. PhD student. 🇹🇳➡️🇺🇸
306K Followers 284 FollowingKaggle is the largest global AI community of developers, researchers, and enthusiasts who compete, collaborate, and benchmark what's next in AI.
54K Followers 2K FollowingPioneers in generative AI. chat/image/video/music. you own outputs. $4.99/mo for DeepAI Pro
cool research at @arxiv_daily
For support email [email protected]
59 Followers 197 FollowingI am a Shark. I have Shark playing cards. I have juggling clubs. I don't have juggling sharks. I know more about Canada than I need to. You'll never feel happy.
546 Followers 408 FollowingCo-founder, CEO of Talus Bio
illuminating the regulome | controlling the genome | drugging the 'undruggable'
#proteomics, #teammassspec, #chembio