Machine Learning

1755 readers
6 users here now

founded 4 years ago
MODERATORS
51
 
 

Hello Machine Learning Community,

The intention of this post is to replicate a similar tradition from R/machinelearning and to trigger engagement. This post will be created weekly.

What are you reading this week and any thoughts to share?

52
 
 

When I train my PyTorch Lightning model on two GPUs on jupyter lab with strategy="ddp_notebook", only two CPUs are used and their usages are 100%. How can I overcome this CPU bottleneck?

Edit: I tested with PyTorchProfiler and it was because of old ssds used on the server

53
54
55
56
57
58
 
 

Hello Machine Learning Community,

The intention of this post is to replicate a similar tradition from R/machinelearning and to trigger engagement. This post will be created weekly.

What are you reading this week and any thoughts to share?

59
 
 
60
61
62
 
 

Hello Machine Learning Community,

The intention of this post is to replicate a similar tradition from R/machinelearning and to trigger engagement. This post will be created weekly.

What are you reading this week and any thoughts to share?

63
 
 

I was looking through papers that combine LLMs and RL and this was pretty fascinating and the citations are perfect for continuing my search.

64
65
66
 
 

Hello Machine Learning Community,

The intention of this post is to replicate a similar tradition from R/machinelearning and to trigger engagement. This post will be created weekly.

What are you reading this week and any thought to share on it ?

67
68
69
 
 

I'd love to know what others are reading, why they think it's awesome (or not). In general, get an exposure to other sub genres of ML. Most of the papers I read are in the computer vision domain cause of work so I'd appreciate reading more about others.

So...

  1. Are you all interested in such a post ?
  2. If yes, which day of the week ?
70
71
 
 

Great series on machine learning. Posting for anyone interested in more of the details on the AI's and LLM's and how they're built/trained.

72
 
 

TLDR of Stability-AI's Paper:

Summary: The document discusses the advancements and limitations of the Stable Diffusion (SDXL) model for text-to-image synthesis. SDXL shows significant improvements in synthesized image quality, prompt adherence, and composition. However, it also has limitations such as challenges in synthesizing intricate structures like human hands, achieving perfect photorealism, addressing biases, mitigating concept bleeding, and improving text rendering. The document also compares SDXL with Midjourney v5.1, where SDXL shows a slight preference in terms of prompt adherence. The document concludes with suggestions for future improvements.

Key Takeaways:

  1. SDXL outperforms or is statistically equal to Midjourney V5.1 in 7 out of 10 categories.
  2. SDXL does not achieve better FID scores than the previous SD versions. This suggests the need for additional quantitative performance scores, specifically for text-to-image foundation models.
  3. SDXL outperforms Midjourney V5.1 in all but two categories in the user preference comparison.
  4. The model may encounter challenges when synthesizing intricate structures, such as human hands.
  5. The model does not attain perfect photorealism. Certain nuances, such as subtle lighting effects or minute texture variations, may still be absent or less faithfully represented in the generated images.
  6. The model’s training process heavily relies on large-scale datasets, which can inadvertently introduce social and racial biases.
  7. The model may exhibit a phenomenon known as “concept bleeding” where distinct visual elements unintentionally merge or overlap.
  8. The model encounters difficulties when rendering long, legible text.
  9. Future work should investigate ways to provide a single stage of equal or better quality, improve text synthesis, enable scaling to much larger transformer-dominated architectures, decrease the compute needed for inference, and increase sampling speed.
73
74
 
 

I work with machine learning tasks daily, both as an ML researcher and as a hobby. The difference between what I can do at work and at home is significant - an A40 at work can do far more than the 3080 I have at home. This obviously makes sense, given the massively increased price point.

However, what I find odd is how there are no consumer level server GPUs targeted towards ML on the market. The A40 is not just a scaled up consumer GPU, and with machine learning growing as a hobby, consumer and enthusiast-level server GPUs are a surprising market gap.

75
 
 

https://public.dhe.ibm.com/ibmdl/export/pub/software/server/ibm-ai/conda/#/

On the face of it, the ability to run models larger than GPU memory would seem to be extremely valuable. Why did they give up? Not everyone has an 80GB GPU.

Was the performance too slow?

view more: ‹ prev next ›