• home
    • members
    • about us
    • books
  • Lectures
    • Deep Learning基礎講座
    • Deep Learning応用講座
    • Deep Learning実践開発講座
    • Deep Learning for NLP講座
    • Deep Learning Day
  • DL Seminars
  • DL Hacks
  • members
  • news
  • contact
    • 講座申し込みのよくある質問
    • contact
  • facebook
  • Twitter
  • English
    • 日本語 日本語
    • English English
Skip to content
Logo Deep Learning JP Discover the Gradient
  • home
    • members
    • about us
    • books
  • Lectures
    • Deep Learning基礎講座
    • Deep Learning応用講座
    •  Deep Learning実践開発講座
    • Deep Learning for NLP講座
    • Deep Learning Day
  • DL Seminars
  • DL Hacks
  • members
  • news
  • contact
    • 講座申し込みのよくある質問
    • contact
  • facebook
  • Twitter
  • English
    • 日本語 日本語
    • English English

Category:

Do Differentiable Simulators Give Better Policy Gradients?

【DL輪読会】 Do Differentiable Simulators Give Better Policy Gradients? by @DeepLearning20 more

read more

DuQuant: Distributing Outliers via Dual Transformation Makes Stronger Quantized LLMs

【DL輪読会】DuQuant: Distributing Outliers via Dual Transformation Makes Stronger Quantize more

read more

1 RomanSetu: Efficiently unlocking multilingual capabilities of Large Language Models via Romanization (ACL 2024) + An Empirical Comparison of Vocabulary Expansion and Initialization Approaches for Language Models (2024)

【DL輪読会】RomanSetu: Efficiently unlocking multilingual capabilities of Large Language M more

read more

Denoising Diffusion Bridge Models

【DL輪読会】Denoising Diffusion Bridge Models by @DeepLearning2023

read more

Classifier-Free Guidance is a Predictor-Corrector

【DL輪読会】Classifier-Free Guidance is a Predictor-Corrector by @DeepLearning2023

read more

Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models on a Synthetic Task

【拡散モデル勉強会】Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models more

read more

Foundation Models for 3D Scene Understanding

Foundation Models for 3D Scene Understanding by @DeepLearning2023

read more

DiG: Scalable and Efficient Diffusion Models with Gated Linear Attention

<script async class=”docswell-embed” src=”https://bcdn.docswell. more

read more

Lory: Fully Differentiable Mixture-of-Experts forAutoregressive Language Model Pre-training

Sorry, this entry is only available in Japanese. For the sake of viewer convenience, more

read more

What Do Language Models Learn in Context? The Structured Task Hypothesis

What Do Language Models Learn in Context? The Structured Task Hypothesis by @DeepLear more

read more
7 / 98« 先頭«...56789...203040...»最後 »
Copyright © 2017 Deep Learning JP. All Rights Reserved.