datascientology | Образование

Telegram-канал datascientology - Data Scientology

1234

Hot data science related posts every hour. Chat: https://telegram.me/r_channels Contacts: @lgyanf

Подписаться на канал

Data Scientology

101. @r_DeepFriedMemes
102. @chessmemes
103. @r_dontdeadopeninside
104. @r_AskReddit
105. @r_kerala
106. @r_AxieInfinity
107. @r_thinkpadsforsale
108. @foxgirls_hot
109. @ImaginaryPics
110. @NFL_reddit
111. @r_etymology
112. @reddit2telegram
113. @AzurLane_sub
114. @rekabufeed
115. @r_kendricklamar
116. @WTF_PICTURES
117. @r_antiwork
118. @RedditCats
119. @redditshortfilms
120. @r_plsnobulli
121. @legalcatadvice
122. @r_nvidia
123. @YoutubeCompendium
124. @NikonBackup
125. @catmemes_reddit
126. @coolguides
127. @rmallubabes
128. @PoliticalHumor
129. @programmingreddit
130. @rsoccerbetting
131. @r_BokuNoMetaAcademia
132. @reddit_trackballs
133. @r_ExpandDong
134. @vfxbackup
135. @RedditGames
136. @r_propagandaposters
137. @r_Unity3D
138. @r_pubgmobile
139. @RussianIsSoHard
140. @Rattit
141. @r_DetroitPistons
142. @r_catastrophicfailure
143. @anime_bikini_waifus
144. @Idiots_In_Cars
145. @r_bapcsalescanada
146. @r_minecraft
147. @r_teenagers
148. @r_raspberry_pi
149. @r_one_punch_man
150. @r_blueteamsec
151. @JEENEETardsReddit
152. @r_traumacore
153. @r_illegallysmolcats
154. @r_chemicalreactiongifs
155. @r_cursed
156. @r_dndgreentext
157. @r_ContraPoints
158. @r_terraria
159. @reddit_elm
160. @Next_Level_Skills
161. @The100backup
162. @r_hackintosh
163. @gameofthronesbackup
164. @r_avatar_memes
165. @failures_of_capitalism
166. @r_okbuddyretard
167. @g4m3savisos
168. @thefalconandthews_reddit
169. @r_k12sysadmin
170. @churchoftohsaka
171. @r_CoolGithubProjects
172. @r_streetwear
173. @CallOfDutyMobile_reddit
174. @r_versus
175. @ImaginationExplorer
176. @one_piece_topic
177. @r_VirginVsChad
178. @r_manga2
179. @r_creepy
180. @UNBGBBIIVCHIDCTIICBG
181. @EliteDanger0us
182. @r_malazan
183. @kstxi
184. @r_okbuddychicanery
185. @r_BikiniBottomTwitter
186. @r_WritingPrompts
187. @SubredditMix
188. @rareinsults
189. @blackpeopletweets
190. @r_okbuddyrintard
191. @r_funnystories
192. @saber_fgo
193. @r_outerwilds
194. @r_hololive
195. @brasildob
196. @Dreamcatcher_reddit
197. @rJackSucksAtLife
198. @oddly_satisfy
199. @instant_regret
200. @imaginary_maps

Читать полностью…

Data Scientology

🎂🎂🎂🎂🎂🎂🎂
🎁 Today @datascientology is 7 years old.
🎉 Congratulations! 🎈

Читать полностью…

Data Scientology

Serverless development experience for embedded computer vision

I recently published the version 1.0 of the Pipeless framework. It provides the development experience of serverless web frameworks to create computer vision applications that run directly on devices.

What that means is you provide some Python functions and those are executed when there is a new video frame. The framework manages everything for you including parallelization, streams management, executing your functions when they have to be executed, etc. You can run it in your devices and provide input streams via a CLI or REST API. It supports multi-stream processing, dynamic stream configuration, ships some inference runtimes so you just need to provide a model, and a bunch of other cool features.

It is working at a very decent performance. I have reached real-time 15 FPS in a CPU of 4 cores with a YOLO model. When using GPU it more than doubles that.

If someone is interested I really appreciate feedback!

You can find the repo here: https://github.com/pipeless-ai/pipeless

/r/computervision
https://redd.it/17vemga

Читать полностью…

Data Scientology

D What AI topics are you curious about but rarely see in the spotlight?

I'm a data engineer who somehow ended up as a software developer. So many of my friends are working now with the OpenAI api to add generative capabilities to their product, but they lack A LOT of context when it comes to how LLMs actually works.

This is why I started writing popular-science style articles that unpack AI concepts for software developers working on real-world application. It started kind of slow, honestly I wrote a bit too "brainy" for them, but now I've found a voice that resonance with this audience much better and I want to ramp up my writing cadence.

I would love to hear your thoughts about what concepts I should write about next?
What get you excited and you find hard to explain to someone with a different background?

/r/MachineLearning
https://redd.it/17riznw

Читать полностью…

Data Scientology

R Idempotent Generative Network

Paper: https://arxiv.org/abs/2311.01462

Blog: https://assafshocher.github.io/IGN/

Abstract:

>We propose a new approach for generative modeling based on training a neural network to be idempotent. An idempotent operator is one that can be applied sequentially without changing the result beyond the initial application, namely f(f(z))=f(z). The proposed model f is trained to map a source distribution (e.g, Gaussian noise) to a target distribution (e.g. realistic images) using the following objectives: (1) Instances from the target distribution should map to themselves, namely f(x)=x. We define the target manifold as the set of all instances that f maps to themselves. (2) Instances that form the source distribution should map onto the defined target manifold. This is achieved by optimizing the idempotence term, f(f(z))=f(z) which encourages the range of f(z) to be on the target manifold. Under ideal assumptions such a process provably converges to the target distribution. This strategy results in a model capable of generating an output in one step, maintaining a consistent latent space, while also allowing sequential applications for refinement. Additionally, we find that by processing inputs from both target and source distributions, the model adeptly projects corrupted or modified data back to the target manifold. This work is a first step towards a ``global projector'' that enables projecting any input into a target data distribution.

​

/r/MachineLearning
https://redd.it/17otzfw

Читать полностью…

Data Scientology

What’s your responsibility as computer vision developer?

I feel like I’m not doing computer vision. I have started a project at my current organization where I am building the defect detection system from scratch. I am mainly spending my time collecting the dataset, labeling them and trying various models. My team uses highly accurate available models from AWS Rekognition and Roboflow which is one click training process. I feel like anyone can collect, label and test the models.

/r/computervision
https://redd.it/17kssfp

Читать полностью…

Data Scientology

imops - ultra fast classical CV algorithms for Python

Hi everyone!

tl/dr: imops is a collection of carefully optimized CV algorithms for numpy arrays of any dimension!

I work in medical imaging, mostly with 3D CT/MRI. This is a pretty computational heavy field with a focus on near-real time processing.

To my surprise, many of the CV algorithms from scipy and skimage are painfully slow. We reimplemented some of them in Cython and added support for arrays of any dimension.

You can find the project here, and the benchmarks section contains a comparison with scipy/skimage counterparts.

If you're interested in contributing, or would like to see another function implemented, don't hesitate to open a PR or create an issue!

/r/computervision
https://redd.it/17h0i9t

Читать полностью…

Data Scientology

D Is Computer Vision dead? - “Quo Vadis, Computer Vision?”

In ICCV23, several top notch researchers shared their insights (in a workshop called “Quo Vadis, Computer Vision?”) wrt the current state of Computer Vision, especially in light of the meteoric raise of LLMs. Has CV stalled? Is CV dead?

E.g.MIT’s professor Bill Freeman, has some interesting points on foundation models: “FM aren’t fundamental, therefore not stable". Jitendra Malik argues "video can describe the world better than text."

/r/MachineLearning
https://redd.it/17eak3w

Читать полностью…

Data Scientology

What should I self-study specifically to become hirable in the NLP/machine learning field

Hi, I have majors in cognitive science, linguistics, philosophy, and a minor in computer science. I know python, java, and SQL at the leetcode/interview level. I have studied algebra, linear algebra, calculus, differential equations. I would love to apply my linguistics knowledge in a tech job (I have not had a tech job at all before). However, I did not get the chance to study NLP or machine learning in college, so I feel like I do not know how to bridge the gap between these two disciplines. What should I do or study to know what I am doing, to be able to get an entry-level NLP position?

/r/LanguageTechnology
https://redd.it/16mrr90

Читать полностью…

Data Scientology

Seeking learning resources that go deep into NLP foundations and which target advanced-intermediate technical learners

# TL;DR:

Semi-experienced MLE seeking to deepen my knowledge of the modeling side of NLP. Can you recommend any courses or other resources to pursue for this?

Ideally I'd like resources which target advanced-intermediate practitioners and which spend only minimal time on theoretical linguistics concepts (e.g., "What is syntax?"; "What is a morpheme?"; "What is distributional semantics?" - Less because that stuff is unimportant, and more because I already know it inside and out.)

------------

# My brief background

Theoretical linguistics grad here. Several years ago and many years out of school, I set off to STEM-up and become a machine learning engineer (MLE) in NLP. I spent a few years hardcore self-studying math, programming, ML/DL, and NLP. In the end I succeeded, learning just enough to get hired as an NLP MLE using only web-based resources and a few PDFs.

I've been in the role for 3-4 years now, and in the interim I have learned a TON of practical skills about SWE and DevOps that I hadn't learned while self-studying the theory. However, my knowledge of ML/DL/NLP theory hasn't actually grown much. This is mostly because where I work, the modeling is left to PhD'ed researchers, not as much the engineers, and not at all the junior engineers like me. So I'd like to get back into learning that side of things.

Because I prioritized breadth over depth during that initial self-study period, I'm passingly familiar with a fair number of NLP tasks and techniques. But I am a master of none. So on this second pass, because now I know the basics and can "speak the lingo", I'd like to go deep, especially on the foundations on NLP.

# What I'm hoping to get from this post

With this background, can anyone recommend (ideally) courses, books, blogs, or other resources for learning things such as the following?

- foundational NLP tasks (e.g., POS-tagging, dependency parsing, NER, sentiment analysis, summarization) and associated popular models/approaches
- probability theory for NLP
- traditional/non-deep NLP
- sequence classification
- clustering/unsupervised methods for NLP
- multitask learning in NLP


# Why? "Who cares?"

At this point, someone may ask:

> You've already "made it" as an MLE, and modeling clearly is not required. So who cares?

I would retort that as one advances from junior to mid to senior and beyond, "hey I can write great code" should gradually take a backseat to "hey I can make informed decisions about what data and models to pursue given the goals and constraints of a specific business case."

I really want to be able to reason about and make pragmatic arguments like "Well, business case A can be re-conceptualized as a kind of question-answering task (random example), therefore it would be reasonable to start with model B and optimize a cost function C with optimizer D. For that, we'd want to collect at least E examples of data type F, and run some model experiments on G cloud resource, ..." Etc. etc etc. Right now I could follow along if a senior engineer came up with such an argument, then probably execute the work mostly on my own. But I would struggle to come up with and then stand behind the argument myself.

I recently had my first technical interviews (for senior level), and while I did well in the coding rounds, I really felt my deficiencies when answering "What would you do in scenario X, and why?" type questions and talking about the nitty gritty of model architectures. My responses were often just like "Um, I'd throw a transformer model at it", in part because that's 90% of what we do where I work, but also because I lack experience which makes me tend towards "when all you have is a hammer everything looks like a nail" style thinking. Hence, this post.

Anyway, enough ramble. I really look forward to any suggestions I receive. Thanks in advance.

/r/LanguageTechnology
https://redd.it/174k2wv

Читать полностью…

Data Scientology

Stable validation curves on NLP project with BERT

/r/deeplearning
https://redd.it/16w3h3f

Читать полностью…

Data Scientology

Meta Unfolds a 'Universe of AI' Across Instagram, Facebook, and WhatsApp

Meta has unveiled colossal AI updates peppered across its platform that would fundamentally alter user experiences on Instagram, Facebook, and WhatsApp, opening up a "universe of AI" solutions.

For the latest advancements in AI, look here first.

https://preview.redd.it/6od0fkjtp1rb1.png?width=2048&format=png&auto=webp&s=e424d2cd2e614728123005b10431c2c13e780871

Spearheading the AI Universe - Meta AI Chatbot

The “advanced conversational assistant” is set to enhance Messenger, WhatsApp, and Instagram services and will be incorporated into upcoming Ray-Ban Meta smart glasses and Quest 3.
Real-time information capabilities have been bolstered through a partnership with Microsoft Bing, and image generation is powered by a new model, Emu.

A Galaxy of AI Personalities

Meta rolled out 28 AIs in beta, featuring sterling personas such as Snoop Dogg, Tom Brady, Kendall Jenner, and Naomi Osaka, thus amplifying the interactivity quotient.

AI Studio - Empowering Businesses

The AI Studio Platform is equipped to enable businesses to build AI chatbots for messaging services on Facebook, Instagram, and Messenger.
Also, Meta will provide a sandbox tool in the upcoming year for users to experiment with creating their own AI.

Generative AI Stickers - A New Co-creating Experience

AI editing tools will allow users to edit images and co-create content with friends.
The tool uses Llama 2 and the new image generation model, Emu, to convert text prompts into stickers in seconds.

Ray-Ban Smart Glasses with Meta AI

The Ray-Ban smart glasses are equipped with Meta AI, allowing users to receive information, incite creativity, and manage the glasses using just their voice.

(source)

P.S. If you like this kind of analysis, I write a free newsletter with the latest and most impactful news in AI. Professionals from Google, Meta, and OpenAI read it daily.

/r/deeplearning
https://redd.it/16uow1h

Читать полностью…

Data Scientology

Number of images for YOLO.

/r/computervision
https://redd.it/16plvwp

Читать полностью…

Data Scientology

How do Large Language Models compare to NLP toolkits for NLP tasks?

I need to do some NLP on text in a number of different languages (English, Spanish, Russian etc). I've experimented using spaCy, stanza and NLTK, as well as some LLMs like ChatGPT, Bard, LLaMa 2 and GPT-4, to do things like lemmatization and POS tagging.

In my experimentation, GPT-4 with adequate prompting outperformed everything else in every language. I wasn't able to spot any errors.

The other LLMs were more or less on par with NLP toolkits: LLMs were a bit more robust to imperfections in the input strings (typos, weird punctuation etc), but were more likely to make very simple mistakes too.

​

Have you guys tried to use LLMs for NLP?

Can you confirm my experimental results, or did you get a different outcome?

Is anyone trying to take advantage of the power of LLMs for these tasks? For instance, is anyone trying to extract NLP features from the insides of models like LLaMa 2?

/r/LanguageTechnology
https://redd.it/16gtrk4

Читать полностью…

Data Scientology

this

/r/MachineLearning
https://redd.it/16ij18f

Читать полностью…

Data Scientology

Other @reddit2telegram channels powered by @r_channels:
01. @r_greenandpleasant
02. @r_photoshopbattles
03. @r_apple
04. @reddit_OSHA
05. @r_moviequotes
06. @PrivacyGuides
07. @reddit_lego
08. @r_nijisanji
09. @facepalmers
10. @rhyderabad
11. @sffpc
12. @reddit_all
13. @r_combatfootage
14. @r_nosafetysmokingfirst
15. @manpill
16. @r_valorant
17. @rshittymoviedetails
18. @r_mapporn
19. @food_from_reddit
20. @r_changemyview
21. @r_smugs
22. @r_freegamefindings
23. @r_Davie504
24. @r_neovim
25. @aesthetic_waifu_wallpapers
26. @r_adporn
27. @r_fantasy
28. @r_indianmemes
29. @Awwducational
30. @RealRacing3TG
31. @r_videomemes
32. @DongistanSub
33. @r_BigAnimeTiddies
34. @R_MildlyPenis
35. @r_animegifs
36. @bestoftweets
37. @lostbackup
38. @r_League_Of_Memes
39. @r_twinpeaks
40. @r_vexillology
41. @r_educationalgifs
42. @r_magiarecord
43. @r_bakchodi
44. @r_movieclub
45. @GameplayMation
46. @quotesporn
47. @r_apexlegends
48. @r_me_irl
49. @r_formuladank
50. @soccerx
51. @harrypotterbackup
52. @rselfie
53. @Boku_No_Hero_Academia_Topic
54. @Emulationx
55. @r_devilmaycry
56. @moonshotcryptos
57. @r_Sino
58. @AlternateReality
59. @BeautifulFemalesbackup
60. @r_wholesome
61. @rickandmorty_en
62. @r_turkeyjerky
63. @r_ComedyCemetery
64. @MemeArea
65. @Fgrandorder
66. @SkinnyWithAbsbackup
67. @MarbleRacing
68. @Unexpected_Reddit
69. @denpasong
70. @r_algotrading
71. @r_animememe
72. @r_leftistvexillology
73. @r_gtaonline
74. @r_wow
75. @aapexlegends_game
76. @r_cutelittlefangs
77. @r_indiaa
78. @r_thinkpad
79. @r_churchofemma
80. @news_reddit
81. @worldnewsvideo
82. @r_InternetIsBeautiful
83. @r_ilMasseo
84. @r_LeagueOfLegends
85. @hololive_yuri
86. @vtuber_en
87. @r_WatchPeopleDieInside
88. @r_bangalore
89. @InstaReality
90. @r_witcher3
91. @r_indiandankmemes
92. @GGPoE
93. @arkotonog
94. @r_jacksepticeye
95. @r_houkai3rd
96. @r_science
97. @r_kochin
98. @r_crackwatch
99. @r_Gintama
100. @JojosBizarreShitposts

Читать полностью…

Data Scientology

Have a live conversation about a basketball game with GPT4V, Whisper, TTS

/r/computervision
https://redd.it/17ywiwp

Читать полностью…

Data Scientology

Start with Large Language Models (LLMs) in 2023

This is a complete guide to start and improve your LLM skills in 2023 without an advanced background in the field and stay up-to-date with the latest news and state-of-the-art techniques!

The complete article: https://www.louisbouchard.ai/from-zero-to-hero-with-llms/

All the links on GitHub: https://github.com/louisfb01/start-llms

Artificial is a fantastic field, and so are language models like GPT-4, Claude..., but it goes extremely fast. Don't miss out on the most important and exciting news by joining great communities, people, newsletters, and more you can all find in this guide!

This guide is intended for anyone with a small background in programming and machine learning. Simple python knowledge is enough to get you started. There is no specific order to follow, but a classic path would be from top to bottom. If you don't like reading books, skip it, if you don't want to follow an online course, you can skip it as well. There is not a single way to become a "LLM expert" and with motivation, you can absolutely achieve it.

/r/deeplearning
https://redd.it/17qo9lt

Читать полностью…

Data Scientology

how well do NLP masters prepare you for industry jobs?

Hey everyone!

I am currently a second-year student enrolled in a linguistics program but I am taking some NLP courses (python programming, machine learning, neural networks, machine translation, databases, maybe statistics). I have a great interest in phonetics & speech technology and am therefore looking into Edinburgh's speech and language processing master as well as other one-year more general NLP masters. However, with all of this I am just unsure if a one-year masters, especially a highly specialised one such as edinburgh's, will sufficiently prepare me for an industry job right after graduation? Should I be looking more into two-years programs such as the ones in Germany?

I would greatly appreciate any input!

/r/LanguageTechnology
https://redd.it/17nm5i9

Читать полностью…

Data Scientology

Part 1: Building Vision Transformer from Scratch: A PyTorch Deep Dive

I've just published the first installment of my Vision Transformer series article. find the full functional example of colab link given in the article
Part 1: Building Vision Transformer from Scratch: A PyTorch Deep Dive Plus a Teaser on LORA for Part 2

pashashaik/part-1-building-vision-transformers-from-scratch-a-pytorch-deep-dive-plus-a-teaser-on-lora-for-beef0f3aef5c">pashashaik/part-1-building-vision-transformers-from-scratch-a-pytorch-deep-dive-plus-a-teaser-on-lora-for-beef0f3aef5c" rel="nofollow">https://medium.com/@pashashaik/part-1-building-vision-transformers-from-scratch-a-pytorch-deep-dive-plus-a-teaser-on-lora-for-beef0f3aef5c

/r/deeplearning
https://redd.it/17m6cix

Читать полностью…

Data Scientology

Give a few prompts to A.I. and it can now generate videos such as this...
https://vimeo.com/877454859

/r/LanguageTechnology
https://redd.it/17g7lfu

Читать полностью…

Data Scientology

Is it useful to take a statistics class for computer vision?

Apologies for inundating the subreddit with questions about courses lol. Am wondering if it is useful somehow though, or only marginally (since stats helps with ML, which helps with computer vision).

Also, is it more useful to know in context of a research/academia setting, or in a industrial setting?

/r/computervision
https://redd.it/17cp5d5

Читать полностью…

Data Scientology

Is it better to take a class on 3D understanding, or a course on the physics of visual appearance?

The former course talks about: Explicit, Implicit, and Neural 3D Representations, Differentiable Rendering, Single-view 3D Prediction: Objects, Scenes, and Humans, Neural Rendering, Multi-view 3D Inference: Radiance Fields, Multi-plane Images, Implicit Surfaces, etc., Generative 3D Models, Shape Abstraction, Mesh and Point cloud processing.

The second course talks about more physics and optics stuff, like principles pf photometry, light fields, reflection, refraction, polarization, caustics, lighting and shadows, BRDFs, vision in bad weather, and applications in aerial, underwater, medical, and microscopic imaging.

At this point, I think I'm interested in biomedical applications of computer vision, so I am leaning towards the second course. However, I don't know enough about the job market out there for computer vision -- it would seem to me that self-driving cars and all would prefer the former course. Furthermore, I wonder if that course is better for expanding my understanding, or the second one. I know that I should pick based on what I want to do, not what's more popular necessarily -- but I also don't quite have that figured out either. I just feel that I don't find generative AI my type of thing (despite it being a big deal lol).

I also have a background in classical signal processing, electromagnetism, etc. being an EE so I thought maybe the 2nd course would complement my background more.

Any advice is greatly appreciated!

/r/computervision
https://redd.it/17bsmnh

Читать полностью…

Data Scientology

Context aware chunking with LLM


I'm working on an embedding and recalll project.

My database is made mainly on a small amount of selected textbooks. With my current chunking strategy, however, the recall does not perform very well since lots of info are lost during the chunking process. I've tried everything... Even with a huge percentage of overlap and using the text separators, lots of info are missing. Also, I tried with lots of methods to generate the text that I use as query: the original question, rephrased (by llm) question or a generic answer generated by LLM. I also tried some kind of keyword or "key phrases ", but as I can see the problem is in the chunking process, not in the query generations.

I then tried to use openai api to chunk the file: the results are amazing... Ok, i had to do a lots of "prompt refinement", but the result is worth it. I mainly used Gpt-3.5-turbo-16k
(obviously gpt4 is best, but damn is expensive with long context. Also text-davinci-003 and it's edit version outperform gpt3.5, but they have only 4k context and are more expensive than 3.5 turbo)

Also, I used the llm to add a series of info and keywords to the Metadata.
Anyway, as a student, that is not economically sustainable for me.

I've seen that llama models are quite able to do that task if used with really low temp and top P, but 7 (and I think even 13B) are not enough to have a an acceptable reliability on the output.

Anyway, I can't run more than a 7B q4 on my hardware.
I've made some research and I've found that replicate could be a good resources, but it doesn't have any model that have more than 4k of context length. The price to push a custom model is too much for me.

Someone have some advice for me? There is some project that is doing something similar? Also, there is some fine tuned llama that is tuned as "edit" model and not "complete" or chat?

Thanks in advance for any kind of answers.

/r/LanguageTechnology
https://redd.it/171r2c1

Читать полностью…

Data Scientology

Why everyone is asking about C++?

When I look at the job posts I see a lot of C++ requirements for AI/DL/ML related jobs.

I assume this is to create optimized models. However when I check online I couldn't see any specific benefit of using C/C++ over python.

When de they plan to use C/C++ and for what? I checked some benchmark comparsions and they're very similar etiher. Furthermore can't we use cython instead of C/C++ anyways?

Do you have any ideas about this?

/r/deeplearning
https://redd.it/16zw8ct

Читать полностью…

Data Scientology

“Decoder-only” Transformer models still have an encoder…right? Otherwise how do they “understand” a prompt?

The original transformer model consisted of both encoder and decoder stages. Since that time, people have created encoder-only models, like BERT, which have no decoder at all and so function well as base models for downstream NLP tasks that require rich representations.

Now we also have lots of “decoder-only“ models, such as GPT-*. These models perform well at creative text generation (though I don’t quite understand how or why).

But in many (all?) use cases of text generation, you start with a prompt. Like the user could ask a question, or describe what it wants the model to do, and the model generates a corresponding response.

If the model’s architecture is truly decoder-only, by what mechanism does it consume the prompt text? It seems like that should be the role of the encoder, to embed the prompt into a representation the model can work with and thereby prime the model to generate the right response?

So yeah, do “decoder-only” models actually have encoders? If so, how are these encoders different from say BERT’s encoder, and why are they called “decoder-only”? If not, then how do the models get access to the prompt?

/r/LanguageTechnology
https://redd.it/16nl811

Читать полностью…

Data Scientology

Is running an open sourced LLM in the cloud via GPU generally cheaper than running a closed sourced LLM?

Assuming using the same cloud service, Is running an open sourced LLM in the cloud via GPU generally cheaper than running a closed sourced LLM? (ie. do we pay a premium when running a closed sourced LLM compared to just running anything on the cloud via GPU?)

One eg. I am thinking of is running Llama 2 13b GPTQ in Microsoft Azure vs. GPT-3.5 Turbo.

I understand there are a lot of parameters to consider (such as choosing which GPU to use in Microsoft Azure etc.), but I am really looking at what’s the cheapest way to run Llama 2 13b GPTQ or a performance-equivalent closed sourced LLM.

/r/LanguageTechnology
https://redd.it/16p6ceo

Читать полностью…

Data Scientology

Why use ONNX with Triton Inference Server? Why use ONNX in general?

Since Triton can support TensorFlow and PyTorch via torchscript. I was wondering why you would want to convert your model to ONNX? Is it simply to use TensorRT?

Also just wanted to know why use ONNX in general? What are the main advantages?

/r/computervision
https://redd.it/16ogz45

Читать полностью…

Data Scientology

iPhone 15 Stereo Imaging

In yesterday’s keynote event Apple released the iPhone 15 pro max. Apparently you can now take 3d images (only available on the iPhone 15 pro). Well, it uses two of its camera lenses to take two images from slightly different angles to perform stereo imaging - obtaining depth.

So I’m sitting here thinking - every iPhone can do that - right? I’m looking at my iPhone 11 Pro Max thinking about writing up a program in iOS that can utilize two lenses and to take a “3d image.”

Sounds like a doable project right? I did stereo imaging and depth estimation projects for one of my classes so I think I can take on the challenge.

/r/computervision
https://redd.it/16ihrtk

Читать полностью…

Data Scientology

D The ML Papers That Rocked Our World (2020-2023)

Hey everyone! 👋

I’ve been on a bit of a deep-dive lately, trying to catch up on all the awesome stuff that’s been happening in the ML space. It got me wondering, from 2020 to 2023, what have been the absolute must-read papers that shook the foundations and got everyone talking?

Whether it’s something that reinvented the wheel in your specific niche or just made waves industry-wide, I wanna hear about it!

I’m curious to see how different the responses will be, and hey, this might even become a go-to list for anyone looking to get the lowdown on the hottest trends and discoveries of the past few years.

Can’t wait to hear your thoughts!

# tl;dr

I decided to aggregate your best suggestions into categories for anyone interested in reading them without searching through the whole comment section in the future.

## Theoretical:

[Neural Networks are Decision Trees](https://arxiv.org/abs/2210.05189)
Cross-Validation Bias due to Unsupervised Preprocessing
[The Forward-Forward Algorithm: Some Preliminary Investigations](https://arxiv.org/abs/2212.13345)
LoRA: Low-Rank Adaptation of Large Language Models (included here as it has applications beyond LLMs)
[Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets](https://arxiv.org/abs/2201.02177)

## Image:

ViT related:
[An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale (ViT)](https://arxiv.org/abs/2010.11929)
Emerging Properties in Self-Supervised Vision Transformers
[Training data-efficient image transformers & distillation through attention](https://arxiv.org/abs/2012.12877v2)
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
[A ConvNet for the 2020s (a CNN that implements several key components that contribute to the performance of Vision Transformers)](https://arxiv.org/abs/2201.03545)
(CLIP) Learning Transferable Visual Models From Natural Language Supervision
Diffusion related:
High-Resolution Image Synthesis with Latent Diffusion Models
[Denoising Diffusion Probabilistic Models (DDPM)](https://arxiv.org/abs/2006.11239)
Classifier-Free Diffusion Guidance
[Taming Transformers for High-Resolution Image Synthesis (VQGAN)](https://arxiv.org/abs/2012.09841)
Segment Anything (SAM)
[DINOv2: Learning Robust Visual Features without Supervision](https://arxiv.org/abs/2304.07193)
Bayesian Flow Networks

## NLP:

[Language Models are Few-Shot Learners (GPT-3)](https://arxiv.org/abs/2005.14165)
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
[Training language models to follow instructions with human feedback](https://arxiv.org/abs/2203.02155)
Training Compute-Optimal Large Language Models (Chinchilla)
[The Flan Collection: Designing Data and Methods for Effective Instruction Tuning](https://arxiv.org/abs/2301.13688)
LLaMA: Open and Efficient Foundation Language Models
[Toolformer: Language Models Can Teach Themselves to Use Tools](https://arxiv.org/abs/2302.04761)

## 3D Rendering:

NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
[Highly accurate protein structure prediction with AlphaFold](https://www.nature.com/articles/s41586-021-03819-2)

## Misc:

Human-level play in the game of Diplomacy by combining language models with strategic reasoning

For a well-made and maintained list of ML resources (not only the newest like here) you can check out

Читать полностью…
Подписаться на канал