본문 바로가기

개발자/인공지능과 인간

오픈소스 LLM 씬의 라이징 스타! 'DeepSeek'

반응형

 

 

기사 참고 링크

 

https://huggingface.co/papers/2401.02954?fbclid=IwZXh0bgNhZW0CMTAAAR2TuWkHpQxmYxI8WQo6SBA0H7mA8t5awYVEqXPSSgDyMWTPEBvFzD9TPks_aem_9F8uxfEHGuOuIxqxY8nFTQ

 

Paper page - DeepSeek LLM: Scaling Open-Source Language Models with Longtermism

Authors: DeepSeek-AI , Xiao Bi , Deli Chen , Guanting Chen , Shanhuang Chen , Chengqi Deng , Honghui Ding , Kai Dong , Qiushi Du , Zhe Fu , Huazuo Gao , Kaige Gao , Wenjun Gao , Ruiqi Ge , Kang Guan , Jianzhong Guo , Guangbo Hao , Zhewen Hao , Ying He , We

huggingface.co

 

 

https://open.substack.com/pub/rosinality/p/deepseek-llm?utm_campaign=post&utm_medium=web

 

DeepSeek LLM

세미나 발제를 위한 자료로 작성했지만 공개할만한 가치가 있을 것 같아 공유합니다.

rosinality.substack.com

 

 

 

https://turingpost.co.kr/p/deepseek-model

 

Topic #10: 오픈소스 LLM 씬의 라이징 스타! 'DeepSeek'을 알ᄋ

빠르게 업계의 선두로 나서는 DeepSeek의 혁신적 모델 개발 과정과 접근법

turingpost.co.kr

 

 

https://composio.dev/blog/notes-on-new-deepseek-v3/

 

Notes on the new Deepseek v3

In this blog we go thorough the new Deepseek v3 and compare it with GPT-4o and 3.5 Sonnet across reasoning, math, coding, & writing tasks.

composio.dev

 

 

https://meetrix.io/articles/deepseekcoder-developer-guide/

 

DeepSeek Coder- Developer Guide

DeepSeek Coder is a cutting-edge series of code language models trained from scratch on 87% code and 13% natural language in English and Chinese, with sizes ranging from 1.3B to 33B versions. Pre-trained on 2 trillion tokens across 80 programming languages

meetrix.io

 

 

 

 

 

 

 

 

 

 

반응형

더욱 좋은 정보를 제공하겠습니다.~ ^^