오픈소스 LLM 씬의 라이징 스타! 'DeepSeek'
기사 참고 링크
Paper page - DeepSeek LLM: Scaling Open-Source Language Models with Longtermism
Authors: DeepSeek-AI , Xiao Bi , Deli Chen , Guanting Chen , Shanhuang Chen , Chengqi Deng , Honghui Ding , Kai Dong , Qiushi Du , Zhe Fu , Huazuo Gao , Kaige Gao , Wenjun Gao , Ruiqi Ge , Kang Guan , Jianzhong Guo , Guangbo Hao , Zhewen Hao , Ying He , We
huggingface.co
https://open.substack.com/pub/rosinality/p/deepseek-llm?utm_campaign=post&utm_medium=web
DeepSeek LLM
세미나 발제를 위한 자료로 작성했지만 공개할만한 가치가 있을 것 같아 공유합니다.
rosinality.substack.com
https://turingpost.co.kr/p/deepseek-model
Topic #10: 오픈소스 LLM 씬의 라이징 스타! 'DeepSeek'을 알ᄋ
빠르게 업계의 선두로 나서는 DeepSeek의 혁신적 모델 개발 과정과 접근법
turingpost.co.kr
https://composio.dev/blog/notes-on-new-deepseek-v3/
Notes on the new Deepseek v3
In this blog we go thorough the new Deepseek v3 and compare it with GPT-4o and 3.5 Sonnet across reasoning, math, coding, & writing tasks.
composio.dev
https://meetrix.io/articles/deepseekcoder-developer-guide/
DeepSeek Coder- Developer Guide
DeepSeek Coder is a cutting-edge series of code language models trained from scratch on 87% code and 13% natural language in English and Chinese, with sizes ranging from 1.3B to 33B versions. Pre-trained on 2 trillion tokens across 80 programming languages
meetrix.io