Hi there, nice to meet you. Follow me at http://belfong.com/tweets
476 stories
·
2 followers

‘The Hardest Working Font in Manhattan’

1 Share

Extraordinary illustrated essay by Marcin Wichary, documenting a typeface — and its long, fascinating, splintered history — that exemplifies the difference between beautiful and pretty. The beauty in Gorton isn’t just in its plainness and hardworking mechanical roots — it’s in the history of the 20th century itself. Gorton became such a part of the world that the bygone world of the previous century imbues how this font makes me feel.

Do yourself a favor and read this one in a comfortable chair, with a tasty slow-sipping beverage, on a screen bigger than a phone. Everything about this piece is exemplary and astounding — the writing, the photography, the depths of research. But most of all, Wichary’s clear passion and appreciation. It’s a love letter.

Read the whole story
Belfong
4 days ago
reply
malaysia
Share this story
Delete

★ Nokia’s Next-Day Internal Competitive Analysis of the Original iPhone Largely Got It Right

1 Share

A few days ago Nokia unveiled their Design Archive at Aalto University in Finland. Fahad X spotted a real gem — an internal confidential slide deck shared within the company the day after Apple had introduced the original iPhone at Macworld Expo in January 2007. To the credit of the team that put this presentation together, they mostly got it:

  • “iPhone touch screen UI may set a new standard of state-of-art: The new user interface may change the standards of the superior user experience for the whole market”. They quote Avi Greengart (a keen and sharp observer): “visually stunning and incredibly responsive”.

  • The on-screen keyboard is mentioned only in passing on a deck titled “Other Great Innovations” as “Predictive, corrective input from on-screen qwerty-keypad”. Most of the existing smartphone makers simply could not believe the iPhone’s touchscreen keyboard would work. Steve Ballmer famously laughed his way into getting sacked from Microsoft over the iPhone’s price and lack of a hardware keyboard. BlackBerry was obsessed with hardware keyboards. This Nokia deck is remarkably open to the idea that Apple was onto something. It does have a bullet point, under “iPhone has the biggest impact on the definition of coolness” (true!), that states: “Even though Steve Jobs emphasized iPhone superiority to ‘Buttons’, it is to be expected that the Consumer QWERTY category will continue to succeed.” But still, this deck is remarkable for acknowledging the potential significance of the iPhone’s keyboard.

  • It’s a swing and a miss regarding third party software: “No mention either of Java support, unusual user input method may be the reason. Lack of Java would shut out a big mass of existing SW.” So close to getting it right. The iPhone’s lack of support for the then-“dominant” Java ME (Micro Edition) platform did shut out all existing mobile software. But all of that software sucked, big time. Sucked to develop, sucked to distribute, sucked to install, sucked to use. Not supporting it was a huge win for the iPhone, just like not supporting Flash Player wound up being a huge win for both the iPhone in particular and the mobile web generally.

  • It’s weird in hindsight that the deck makes mention of Java mobile apps but the only mention of the web is in passing: “Browsing - Safari web browser – (reportedly most advanced ever) Full screen with touch zoom-in functionality.” This speaks to Nokia’s pre-iPhone mindset that mobile platforms were not really internet-first devices but instead were mostly for running shitty software from carriers (for consumers) or corporate IT departments. The iPhone’s excellent day-one version of Mobile Safari, with support for the “real internet”, not the “baby internet” (to use Steve Jobs’s own terms). It’s so easy to get blinded by the way things currently work and to assume they’ll keep working that way.

  • The Nokia team nailed the remarkable and potentially industry-changing nature of Apple’s relationship with Cingular:

    • Cingular got multi-year exclusivity to iPhone. In exchange it gave up to Apple in many respects:
      • Cingular has allowed Apple to launch a device with WLAN and inbuilt services
      • The Yahoo and Google experiences are built in to the Apple phone
      • iTunes synch is done in cradle, no OTA
      • The multi-year exclusivity of the Cingular — Apple arrangement makes one ask if Apple gets a share of data revenues or Cingular iTunes revenues?
      • This may be a start for the whole market to change
    • iPhone will be sold both in Apple stores and Cingular stores
    • This is a marriage of convenience built on mutual distrust and recognition that each has assets the other needs

    Nokia concluded, “Other US operators need desperately something against Cingular and Apple” and advised that Nokia should “work very closely especially with T-Mobile.”

The most prescient line in the whole deck was this, recognizing that the iPhone was a UI breakthrough and that the entirety of Nokia’s offerings were far behind:

Nokia needs to develop the touch UI to fight back. S60 should be focus. Maemo platform is critical strength due to openness. Nokia needs a Chief UI architect to re-energize Nokia’s UI innovation across platforms and businesses.

I’d translate that as the presentation team imploring Nokia’s leadership to recognize that Apple’s primary priority was creating and delivering a great experience, both hardware and software, and Nokia had no such institutional value placed on design quality — on coolness — and would need to reinvent itself to prioritize the same ideals if it was going to compete.

Upton Sinclair’s famed adage is apt, as ever: “It is difficult to get a man to understand something, when his salary depends on his not understanding it.”

Read the whole story
Belfong
33 days ago
reply
malaysia
Share this story
Delete

Trying out QvQ - Qwen's new visual reasoning model

1 Comment and 3 Shares

I thought we were done for major model releases in 2024, but apparently not: Alibaba's Qwen team just dropped the Apache 2.0 licensed Qwen licensed (the license changed) QvQ-72B-Preview, "an experimental research model focusing on enhancing visual reasoning capabilities".

Their blog post is titled QvQ: To See the World with Wisdom - similar flowery language to their QwQ announcement QwQ: Reflect Deeply on the Boundaries of the Unknown a few weeks ago in November.

It's a vision-focused follow-up to QwQ, which I wrote about previousy. QwQ is an impressive openly licensed inference-scaling model: give it a prompt and it will think out loud over many tokens while trying to derive a good answer, similar to OpenAI's o1 and o3 models.

The new QvQ adds vision to the mix. You can try it out on Hugging Face Spaces - it accepts an image and a single prompt and then streams out a very long response where it thinks through the problem you have posed it. There's no option to send a follow-up prompt.

I've tried it out with a bunch of things, with mixed results - but it's really fun seeing how it works through a problem.

My most successful prompt was "Count the pelicans" with this image:

A photograph of four glorious young California Brown Pelicans perched on a dock. The back end of another pelican is visible at the front of the photo.

Here's the full response. Some highlights:

So I've got this picture of some pelicans, and I need to count them. Let me see... First off, I know that pelicans are those big birds with the massive beaks, right? [...]

Let me try counting again, more carefully this time. Starting from the left:

  1. The first pelican is fully visible, standing at the front.
  2. Behind it, slightly to the right, is the second pelican.
  3. Then, further back, there's a third pelican.
  4. And next to that one, there's a fourth pelican.

So, that's four pelicans in total. But, I notice that on the far right of the image, there's part of another bird, maybe a pelican's back or tail feathers, but it's not fully in the frame, so I don't think I should count it as a whole pelican.

Something about the tone of voice it uses, "First off, I know that pelicans are those big birds with the massive beaks, right?", tickles me.

I also tried this on a much harder image - it counted 17 which looks about right to me but honestly I have trouble with that image myself.

I fed in one of the ARC-AGI puzzles that o3 had failed at.

A puzzle with colored squares, where drawing a line between the single blue squares and turning any intersected rectangles blue is clearly the solution.

It produced a very lengthy chain of thought that was almost entirely incorrect, but had some interesting "ideas" in it:

[...] Let me try to think in terms of cellular automata rules, like Conway's Game of Life.

In Game of Life, each cell's state in the next generation is determined by its current state and the states of its eight neighbors.

Perhaps a similar rule applies here.

Finally, I asked it to "Estimate the height of the dinosaur" against this image (which, as it correctly noted, is actually an inflatable dragon):

A large inflatable Christmas decoration of a reindeer riding a dragon - both reindeer and dragon are wearing Christmas hats

I'm not sure this task is even possible given the lack of items in the frame of comparable known height. It had a good go at it though! I particularly enjoyed how it began:

So I've got this inflatable Christmas decoration, and it's a dragon! But not just any dragon—it's a green dragon wearing a Santa hat and a red and white scarf. And get this, there's a reindeer on its back, also wearing a Santa hat and holding onto the dragon's wings. It's pretty festive and colorful, with lots of vibrant greens, reds, and yellows.

The dragon has these big, friendly eyes and a smiling face, which makes it look really cheerful. Its wings are spread out, and it has white claws on its feet. The reindeer is brown with antlers and seems to be enjoying the ride on the dragon's back.

Now, the question is, how tall is this inflatable dragon? I need to estimate its height. Let me see. I'm standing next to it, and I'm about 5 feet 10 inches tall. The dragon is definitely taller than me. Maybe around 8 or 9 feet high? But I should try to be more precise.

I wonder how it decided that its own height was 5 feet 10 inches!

Running QvQ locally

All of my experiments so far have used the hosted Hugging Face demo. I'm hoping to try it out on my own laptop soon - I think it should just fit in 64GB of M2, maybe with a smaller quantization.

Right now the options are:

As a happy user of Ollama's qwq port I'm hoping they add a QvQ release at some point soon as well.

Running QvQ on my Mac using mlx-vlm

Update: I successfully ran QvQ on my 64GB M2 MacBook Pro using Prince Canuma's mlx-vlm via uv run with the following command:

uv run --with 'numpy<2.0' --with mlx-vlm python \
  -m mlx_vlm.generate \
    --model mlx-community/QVQ-72B-Preview-4bit \
    --max-tokens 10000 \
    --temp 0.0 \
    --prompt "describe this" \
    --image pelicans-on-bicycles-veo2.jpg

My image input here was taken from my experiment with Google's Veo 2 video generation model:

A screenshot of the Google VideoFX video generation tool

Here's the full output from running that command, after it first downloaded 38GB to ~/.cache/huggingface/hub/models--mlx-community--QVQ-72B-Preview-4bi:

==========
Image: ['pelicans-on-bicycles-veo2.jpg'] 

Prompt: <|im_start|>system
You are a helpful and harmless assistant. You are Qwen developed by Alibaba. You should think step-by-step.<|im_end|>
<|im_start|>user
describe this<|vision_start|><|image_pad|><|vision_end|><|im_end|>
<|im_start|>assistant

Alright, I've got this "Text to Video" tool to explore. It seems pretty advanced, allowing me to turn text descriptions into actual videos. The interface looks user-friendly, with a dark theme that's easy on the eyes. On the left side, there's a panel where I can input my text prompt. It already has an example filled in: "A pelican riding a bicycle along a coastal path overlooking a harbor."

That sounds interesting! I can see four different video thumbnails on the right side, each showing variations of this scene. One of them even has the pelican wearing a helmet, which adds a humorous touch. Another shows the pelican standing next to a bicycle, almost as if it's ready to hop on and start pedaling.

Below the text input, there are some style options like "photorealistic," "35mm film," "minimal," "sketchy," "handmade," and "abstract." These probably let me choose the visual style of the video. For now, I'll stick with the default settings.

At the bottom of the left panel, there are two buttons: "Start over" and "Create videos." I guess "Create videos" is the one I need to press to generate the videos based on my text prompt.

Let me think about this. I want to see what kind of videos this tool can produce based on my description. The example prompt is already filled in, so maybe I should just go ahead and click "Create videos" to see what happens.

Wait a minute, the prompt mentions a coastal path overlooking a harbor. In the thumbnails, I can see a background with water and boats, which matches the description. It's cool how the tool has interpreted the text into visual elements.

I should also pay attention to the "Visual Prompt" section, which says "View only." That probably means I can't edit it directly, but I can view the generated videos based on this prompt.

Hmm, maybe I should try changing the prompt to see different results. For example, instead of a pelican riding a bicycle, maybe I could describe a different scene. But for now, I'll stick with the given example to understand how it works.

After clicking "Create videos," I expect to see the videos being generated. There might be a loading process, depending on the complexity of the prompt and the tool's processing speed.

Once the videos are ready, I can preview them to see how well they match my description. If I'm not satisfied, I can always go back and adjust the prompt or try different style options.

Overall, this seems like a powerful tool for creating custom videos based on textual descriptions. It could be really useful for storytelling, presentations, or even just for fun, experimenting with different scenes and styles.

I think I'll spend some time exploring the different style options to see how they affect the final videos. For instance, selecting "35mm film" might give the videos a vintage, cinematic look, while "abstract" could produce more artistic and less realistic visuals.

Also, I wonder if there's a way to customize more specific aspects, like the camera angle, lighting, or movement. Maybe after generating the initial videos, there are additional editing options available.

In summary, I'm excited to use this Text to Video tool and see what kind of creative possibilities it opens up. With a bit of practice, I could probably create some impressive visual content just by describing what I want to see.

**Final Answer**

To proceed with generating videos based on the provided text prompt, I need to click on the "Create videos" button. This action will initiate the process of converting the textual description into visual content, resulting in multiple video thumbnails that can be previewed for further use or adjustments.
==========
Prompt: 0.870 tokens-per-sec
Generation: 7.694 tokens-per-sec

The license changed from Apache 2.0 to Qwen

When I wrote this post yesterday the LICENSE file in the Hugging Face repository was Apache 2.0. Just after midnight UTC on 25th December this commit landed updating the QVQ-72B-Preview license file to the Qwen license instead.

This looks to me like they were correcting a mistake, not changing their policy. The README.md for that repository has this block of YAML:

license: other
license_name: qwen

And commits to that README at one point linked to the Qwen2.5-72B-Instruct copy of the Qwen license.

The QwQ model repository continues to list Apache 2.0, which matches the YAML in its README as well.

So it looks to me like the intention is for QvQ and Qwen2.5-72B-Instruct to be Qwen licensed, while QwQ is Apache 2.0.

Tags: python, ai, generative-ai, llms, hugging-face, vision-llms, uv, qwen, mlx, inference-scaling

Read the whole story
samuel
59 days ago
reply
I wonder how large of an M4 Pro or M4 Max I’ll need to run this model locally. Visual reasoning locally means I can throw some interesting problems at it and run it all night. Find needles in a haystack of my own photo library, for instance.
Cambridge, Massachusetts
kglitchy
58 days ago
That's a great idea! What are some use cases for information you would want to gather?
samuel
58 days ago
I want to classify photos so I can more easily make year-end photo books
samuel
58 days ago
Also, I'm about to launch embeddings (for text) on NewsBlur (try it out at staging.newsblur.com - click on the Related buttons in feed titles and at the bottom of stories) and this sort of transformer based classifier would be hugely helpful
kglitchy
57 days ago
Great ideas and I like the related stories idea!
Belfong
57 days ago
reply
malaysia
Share this story
Delete

Coding Font Selection ‘Tournament’

1 Share

Via Jason Snell (back in October), who points first to this thread on Mastodon where a few of us posted about our preferences for the fonts we use for writing, and then describes this fun “tournament” from Typogram that lets you pick your favorite monospaced coding font from 32 choices. One limitation is that the only options are free fonts — some of my favorite monospaced fonts aren’t free and thus aren’t included (e.g. Consolas, Berkeley Mono, or Apple’s SF Mono). Another limitation is that some of the fonts in the tournament just plain suck. But it’s really pretty fun.

It’s also a good thing I procrastinated on linking to this for two months — it’s improved greatly in the weeks since Snell linked to it. The example code is now JavaScript, not CSS, which is a much better baseline for choosing a programming font. And there are some better font choices now.

I highly recommend you disable showing the font names while you play, to avoid any bias toward fonts you already think you have an opinion about. But no matter how many times I play, I always get the same winner: Adobe’s Source Code Pro. My second favorite in this tournament is IBM Plex Mono. The most conspicuous omission: Intel One Mono.

Read the whole story
Belfong
58 days ago
reply
malaysia
Share this story
Delete

A look at Mozilla’s rebrand. “Mozilla’s new logo is a flag symbol...

1 Comment
A look at Mozilla’s rebrand. “Mozilla’s new logo is a flag symbol built from the Mozilla M that comes to life to reveal the company’s iconic Tyrannosaurus Rex symbol and mascot, originally designed by Shepard Fairey.”

💬 Join the discussion on kottke.org

Read the whole story
Belfong
136 days ago
reply
I love it!
malaysia
Share this story
Delete

What happens when solar panels die?

1 Comment and 3 Shares

By the end of 2024, the world will have nearly 2,000 Gigawatts of solar generation capacity in service. Each panel is made of silicon, glass, various polymers, aluminum, copper and an assortment of other metals that capture the sun’s energy. It’s a rule of thumb that, barring damage, a panel will last for up to 30 years before it needs to be replaced. But what happens to all of those raw materials when the current crop of solar panels becomes obsolete? Surely, we’re not just wasting it all, are we?

Received wisdom suggests solar panels last for around 30 years, but that’s not the whole story here. “30 years is our best guess,” explained Garvin Heath of the National Renewable Energy Laboratory (NREL). NREL found there was a higher rate of failures at the start of a panel’s life, often due to manufacturing or installation faults. In midlife, only a handful of panels fail. Then the statistics begin to climb northward the closer to the three decade mark you get but, even so, the number of panels that break are “less than one percent” of the total in operation at that time.

Matt Burnell is the founder of ReSolar, a British startup looking into reusing, repowering and recycling solar panels. As part of his work, Burnell visited a 40,000 panel array solar farm where 200 of the panels were broken during installation. “I took about 50 from that site, tested them to see their value for reuse [and] generation capacity,” he said, most of which were within the “tolerance range of the manufacturer.” Essentially, for the odd crack in the glass or bump on the frame — which may cause problems down the line — the panels were otherwise perfectly functional.

If a panel has survived its birth and installation, then the biggest thing that kills solar panels is the weather. Heath said a common cause is extreme weather events damaging the panel, or even just regular, aggressive weather causing things to degrade. Sadly, once a panel is broken, it’s often not worth the effort to repair.

So panels deemed “broken” during manufacture or installation may still be very capable of making power from the sun. But there are also plenty of panels that are being withdrawn from service after 25 or 30 years, even if they aren't broken in any meaningful sense. There's a fairly simple reason solar farms don't allow these panels to soak up rays until they simply cease to function.

The key issue is efficiency loss, which is when panels aren’t able to generate as much power as they did when first installed. Most solar panels are made with laminated adhesive layers that sit between the glass and the solar cells to hold them together and aid rigidity. Sun exposure can cause those laminated layers to discolor, reducing the amount of light that can reach the cells. That diminishes the energy-generation capacity, which is a problem for large commercial farms.

“Manufacturer's warranty their [solar] modules’ performance for a 30-year period,” explained Garvin Heath. For instance, a maker will pledge that its panels will be at least 80-percent efficient for the bulk of its expected three-decade service life. These warranties give large utility-scale customers confidence in what they’re buying, and at the point that term has expired, it’s often far more cost-effective to simply junk and replace them.

Power grids have a limited number of interconnections, essentially the on-ramp that enables them to push power to the grid. Each interconnection has a hard upper limit in terms of the power it can send, so solar farms need to generate the maximum permitted electricity at all times. “[Even when] they’re working within warranty performance, the opportunity cost of having a module producing [more] power on your interconnection is quite valuable,” said Heath.

ReSolar’s Matt Burnell used an example of a 10 Megawatt solar farm in the UK that had a 15 Megawatt interconnection. “10 years ago, they could only fit 10 megawatts into the space that they had [...] but with newer and more efficient modules, it’s now financially viable for them to strip the asset down and rebuild it.” “You have these big pension funds looking at this from a spreadsheet,” looking for ways to better maximize their investment. The end result is that all of these otherwise fine panels are junked. “When you think about the embedded carbon of bringing [the panels] over [from China]” said Burnell “and then they go into the waste stream [...] seems mad.”

Even if panels could be repaired to full efficiency,it’s not likely solar panel repair shops will be opening in droves. “There’s a serious question around the labor costs of testing and repairing versus just buying a new panel,” said Burnell. He added in another example of panels that had to be taken down to address fire safety legislation, which were similarly at risk of being discarded because the effort to repurpose them was too great. To reduce waste, ReSolar actually wound up collecting and sending on a consignment of those panels to Ukraine for use in a hospital.

Another rule of thumb is that only one in 10 solar panels is recycled, with the remaining nine sent to landfills. There is no standard method for tracking a panel’s eventual destination, and it’s not clear how such a system would be implemented. But there’s a risk landfills are about to be overwhelmed with the volume of panels that’ll be coming down from roofs. The Los Angeles Times, for instance, reported on the coming glut of panels in California after the state’s push to get more solar installed from 2006 onwards.

The legal situation is barely patchwork, with Grist describing things in 2020 as the “wild west,” since only Washington has any sort of mandatory legislation. Decommissioned solar panels are covered by federal solid and hazardous waste rules, dependent on the materials used in their construction. If a panel includes heavy metals like lead and cadmium, then they can’t be sent to a general landfill, lest their poisons leech into the soil. But that often just means those panels are redirected to landfills that are designed to handle specialist waste.

The EPA is, at present, looking at developing rules that would standardize the recycling process for solar panels and lithium batteries. But while there are no federal mandates for recycling, or even tough legislation at the state level, the situation is far from ideal. A small fraction of the panels are actually sent to recycling centers, the rest left to an uncertain fate. As Heath points out, the risk is that while recycling is uneconomical and unavailable, we’ll see huge boneyards of working solar panels, left piled up while the situation changes.

In the UK and Europe, solar panels are covered by the Waste from Electrical and Electronic Equipment directive, or WEEE. The rules oblige supplying companies to collect and recycle discarded panels, or to shoulder the cost for another entity to do so. It means that, hopefully, we won’t see tons more panels being dumped to landfills, but also means it’s often going to be more economical to send working panels to recycling rather than repurposing them.

If you want to free up the raw materials lurking inside a solar panel, then there are two approaches. There’s the mechanical way, in which you can shred the components, which is both simpler and more wasteful: it can recover glass and metal, but little else. Or there are thermal and chemical approaches that seek to separate the components, enabling more of the rarer metals to be recovered.

“Existing recyclers have traditional markets that their economics are built around, so glass recyclers look at a module and say ‘wow, a module is 80 percent glass by weight, I know what to do with that,” said Heath. “With the materials inside, there are more precious metals with higher value,” he said, “but they’re mixed in with the plastic polymer layers [...] which are hard to separate economically.” Consequently, the silicon, silver and copper embedded in the cells are often ground down into bulk and abandoned.

The IEA’s 2024 report on panel recycling looked into how these mechanical methods aren’t great for material qualities. “The outputs of mechanical processing are usually not very pure and better yields of high-quality materials [...] especially silicon and silver, should be targeted,” it said. It added that often these recycling processes aren’t optimized to run solar panels, and so “there is frequently some downgrading of recovered material quality,” hardly a great step on the road to circularity.

It’s also hard to know what goes into a solar panel. “The variation in materials [found in solar panels] is wild,” said ReSolar’s Matt Burnell. The litany of manufacturers don’t yet have any obligation to share their raw material data, although new regulations will change that soon. Until then, it’s difficult for recyclers to know what they’ll be pulling out of the panels they’re looking to process.

As well as recyclers not knowing the composition of the panels, there’s the risk of noxious chemicals being added to expedite some processes. Antoine Chalaux is the general manager of ROSI Solar, a specialist solar panel recycler in France. He talked about the inclusion of chemicals like Teflon and antimony, both of which are toxic and cannot be released into the atmosphere. “We’ve developed our recycling processes to capture [them],” he explained, “but we’re pushing [manufacturers] to use it less [in future].”

Burnell believes that the industry is really at the “very dawn” of solar recycling but is confident that with investment today, solutions will be quickly found in the very near future. “We’ve got this massive lead-in time,” he said “so we know what’s coming onto the market today, and we know what’s coming into the system in 25 to 30 years.” The real ticking clock is for the glut of panels that were installed in the early 2010s that will start entering the waste stream in the next decade.

Right now, ROSI’s processes aren’t as cheap as other recyclers, and Chalaux knows that it can be a problem. “Right now, there’s no economic reason for companies to [recycle with us], but there’s the question of image,” he said. “All of the manufacturers and owners of PV projects want a good story for the end of life for their panels.” The other benefit of this process, however, is to produce high-purity recycled materials that can be used by local manufacturers.

One step toward a more recyclable solar panel might be to eliminate the use of those adhesive polymers in its construction. If a panel could just use sheets of glass with the solar cells sandwiched inside, it would be a lot easier to deconstruct. Not to mention you’d likely get a longer and better performance out of them, since there would be no polymer layers to discolor.

Thankfully, a team from the US National Renewable Energy Laboratory (NREL) has demonstrated that such a product can exist. Rather than gluing the layers together, femtosecond lasers weld the front and back panels of glass to each other. The solar cells are sandwiched inside, held by the bonding of the glass to its sibling, and nothing else. And when the panel eventually reaches its end of life, which may be a lot longer than 30 years, it can just be recycled by shattering the glass.

The project, led by Dr. David Young, says that if the proposals are accepted, we could see a commercial version of the panel within two to three years. He added that the rigidity offered by welding will be just as sturdy and waterproof as panels using polymer layers. Unfortunately, by that point, we’ll have decades upon decades of panels made using the old system that we’ll still need to deal with. And until we get a cost-effective, scalable way to recycle them, the answer to the question ‘What happens to solar panels when they die?’ will be ‘nothing good.’

Read the whole story
Belfong
137 days ago
reply
Useful to note when researching about solar panels in the future.
malaysia
Share this story
Delete
Next Page of Stories