DeepSeekを理解する
Claude
ChatGPT
Gemini
Large Language Model
強化学習
DeepSeekV3 vs Claude-Sonnet vs o1-Mini vs Gemini-ept-1206, tested on real world scenario
https://www.reddit.com/r/LocalLLaMA/comments/1hoe75l/deepseekv3_vs_claudesonnet_vs_o1mini_vs/
https://x.com/deepseek_ai/status/1872242657348710721
Biggest leap forward yet:
⚡ 60 tokens/second (3x faster than V2!)
💪 Enhanced capabilities
🛠 API compatibility intact
🌍 Fully open-source models & papers
https://github.com/deepseek-ai/DeepSeek-V3/blob/main/DeepSeek_V3.pdf
https://github.com/deepseek-ai/DeepSeek-V3
PFN 岡野原さんの考察
https://x.com/hillbig/status/1876397959841186148
安野さん考察
https://x.com/takahiroanno/status/1884260101126816098
Ollamaを活用したDeepSeek-R1のローカルLLM構築ガイド
https://chatgpt-enterprise.jp/blog/ollama-deepseek-r1/
DeepSeek-R1をローカル環境で動かしたらあっさり動いた
https://zenn.dev/showyou/articles/badc504f25a178
DeepSeek-R1
https://github.com/deepseek-ai/DeepSeek-R1/blob/main/DeepSeek_R1.pdf
生成AIでなんかやれと言われる皆さんへ(DeepSeekとばっちり編)
https://anond.hatelabo.jp/20250129194915
How to deploy and fine-tune DeepSeek models on AWS
https://huggingface.co/blog/deepseek-r1-aws
unsloth/DeepSeek-R1-GGUF
https://huggingface.co/unsloth/DeepSeek-R1-GGUF
How has DeepSeek improved the Transformer architecture?
https://epoch.ai/gradient-updates/how-has-deepseek-improved-the-transformer-architecture
Tutorial: Integrating Locally-Run DeepSeek R1 Distilled Llama Model with Cursor
https://www.secondstate.io/articles/deepseek-cursor/
DeepSeek-R1-Distill-Qwen-32B-Japanese
https://huggingface.co/cyberagent/DeepSeek-R1-Distill-Qwen-32B-Japanese
LLMチューニングのための強化学習:GRPO(Group Relative Policy Optimization)
https://horomary.hatenablog.com/entry/2025/01/26/204545
TinyZero
https://github.com/Jiayi-Pan/TinyZero
DeepSeek-R1(deepseek-reasoner)についてまとめ、API から試してみる
https://note.com/schroneko/n/n939f3943f2a6
DeepSeek-R1の論文から読み解く背景技術
https://speakerdeck.com/personabb/deepseek-r1nolun-wen-karadu-mijie-kubei-jing-ji-shu?slide=2