call_end

    • chevron_right

      Asman Malika: Think About Your Audience

      news.movim.eu / PlanetGnome • Yesterday - 12:07 • 3 minutes

    When I started writing this blog, I didn’t fully understand what “think about your audience” really meant. At first, it sounded like advice meant for marketers or professional writers. But over time, I’ve realized it’s one of the most important lessons I’m learning, not just for writing, but for building software and contributing to open source.

    Who I’m Writing (and Building) For

    When I sit down to write, I think about a few people.

    I think about aspiring developers from non-traditional backgrounds, people who didn’t follow a straight path into tech, who might be self-taught, switching careers, or learning in community-driven programs. I think about people who feel like they don’t quite belong in tech yet, and are looking for proof that they do.

    I also think about my past self, about some months ago. Back then, everything felt overwhelming: the tools, the terminology, the imposter syndrome. I remember wishing I could read honest stories from people who were still in the process , not just those who had already “made it.”

    And finally, I think about the open-source community I’m now part of: contributors, maintainers, and users who rely on the software we build.

    Why My Audience Matters to My Work

    Thinking about my audience has changed how I approach my work on Papers.

    Papers isn’t just a codebase, it’s a tool used by researchers, students, and academics to manage references and organize their work. When I think about those users, I stop seeing bugs as abstract issues and start seeing them as real problems that affect real people’s workflows.

    The same applies to documentation. Remembering how confusing things felt when I was a beginner pushes me to write clearer commit messages, better explanations, and more accessible documentation. I’m no longer writing just to “get the task done”. I’m writing so that someone else, maybe a first-time contributor, can understand and build on my work.

    Even this blog is shaped by that mindset. After my first post, someone commented and shared how it resonated with them. That moment reminded me that words can matter just as much as code.

    What My Audience Needs From Me

    I’ve learned that people don’t just want success stories. They want honesty.

    They want to hear about the struggle, the confusion, and the small wins in between. They want proof that non-traditional paths into tech are valid. They want practical lessons they can apply, not just motivation quotes.

    Most of all, they want representation and reassurance. Seeing someone who looks like them, or comes from a similar background, navigating open source and learning in public can make the journey feel possible.

    That’s a responsibility I take seriously.

    How I’ve Adjusted Along the Way

    Because I’m thinking about my audience, I’ve changed how I share my journey.

    I explain things more clearly. I reflect more deeply on what I’m learning instead of just listing achievements. I’m more intentional about connecting my experiences, debugging a feature, reading unfamiliar code, asking questions in the GNOME community, to lessons others can take away.

    Understanding the Papers user base has also influenced how I approach features and fixes. Understanding my blog audience has influenced how I communicate. In both cases, empathy plays a huge role.

    Moving Forward

    Thinking about my audience has taught me that good software and good writing have something in common: they’re built with people in mind.

    As I continue this internship and this blog, I want to keep building tools that are accessible, contributing in ways that lower barriers, and sharing my journey honestly. If even one person reads this and feels more capable, or more encouraged to try, then it’s worth it.

    That’s who I’m writing for. And that’s who I’m building for.

    • chevron_right

      Flathub Blog: What's new in Vorarbeiter

      news.movim.eu / PlanetGnome • 1 days ago - 00:00 • 2 minutes

    It is almost a year since the switch to Vorarbeiter for building and publishing apps. We've made several improvements since then, and it's time to brag about them.

    RunsOn

    In the initial announcement, I mentioned we were using RunsOn , a just-in-time runner provisioning system, to build large apps such as Chromium. Since then, we have fully switched to RunsOn for all builds. Free GitHub runners available to open source projects are heavily overloaded and there are limits on how many concurrent builds can run at a time. With RunsOn, we can request an arbitrary number of threads, memory and disk space, for less than if we were to use paid GitHub runners.

    We also rely more on spot instances, which are even cheaper than the usual on demand machines. The downside is that jobs sometimes get interrupted. To avoid spending too much time on retry ping-pong, builds retried with the special bot, retry command use the on-demand instances from the get-go. The same catch applies to large builds, which are unlikely to finish in time before spot instances are reclaimed.

    The cost breakdown since May 2025 is as follows:

    Cost breakdown

    Once again, we are not actually paying for anything thanks to the AWS credits for open source projects program . Thank you RunsOn team and AWS for making this possible!

    Caching

    Vorarbeiter now supports caching downloads and ccache files between builds. Everything is an OCI image if you are feeling brave enough, and so we are storing the per-app cache with ORAS in GitHub Container Registry.

    This is especially useful for cosmetic rebuilds and minor version bumps, where most of the source code remains the same. Your mileage may vary for anything more complex.

    End-of-life without rebuilding

    One of the Buildbot limitations was that it was difficult to retrofit pull requests marking apps as end-of-life without rebuilding them. Flat-manager itself exposes an API call for this since 2019 but we could not really use it, as apps had to be in a buildable state only to deprecate them.

    Vorarbeiter will now detect that a PR modifies only the end-of-life keys in the flathub.json file, skip test and regular builds, and directly use the flat-manager API to republish the app with the EOL flag set post-merge.

    Web UI

    GitHub's UI isn't really built for a centralized repository building other repositories. My love-hate relationship with Buildbot made me want to have a similar dashboard for Vorarbeiter.

    The new web UI uses PicoCSS and HTMX to provide a tidy table of recent builds. It is unlikely to be particularly interesting to end users, but kinkshaming is not nice, okay? I like to know what's being built and now you can too here .

    Reproducible builds

    We have started testing binary reproducibility of x86_64 builds targetting the stable repository. This is possible thanks to flathub-repro-checker , a tool doing the necessary legwork to recreate the build environment and compare the result of the rebuild with what is published on Flathub.

    While these tests have been running for a while now, we have recently restarted them from scratch after enabling S3 storage for diffoscope artifacts. The current status is on the reproducible builds page .

    Failures are not currently acted on. When we collect more results, we may start to surface them to app maintainers for investigation. We also don't test direct uploads at the moment.

    • chevron_right

      Jussi Pakkanen: How to get banned from Facebook in one simple step

      news.movim.eu / PlanetGnome • 2 days ago - 18:06 • 2 minutes

    I, too, have (or as you can probably guess from the title of this post, had) a Facebook account. I only ever used it for two purposes.

    1. Finding out what friends I rarely see are doing
    2. Getting invites to events
    Facebook has over the years made usage #1 pretty much impossible. My feed contains approximately 1% posts by my friends and 99% ads for image meme "humor" groups whose expected amusement value seems to be approximately the same as punching yourself in the groin.

    Still, every now and then I get a glimpse of a post by the people I actively chose to follow. Specifically a friend was pondering about the behaviour of people who do happy birthday posts on profiles of deceased people. Like, if you have not kept up with someone enough to know that they are dead, why would you feel the need to post congratulations on their profile pages.

    I wrote a reply which is replicated below. It is not accurate as it is a translation and I no longer have access to the original post.

    Some of these might come via recommendations by AI assistants. Maybe in the future AI bots from people who themselves are dead carry on posting birthday congratulations on profiles of other dead people. A sort of a social media for the deceased, if you will.

    Roughly one minute later my account was suspended. Let that be a lesson to you all. Do not mention the Dead Internet Theory , for doing so threatens Facebook's ad revenue and is thus taboo. (A more probable explanation is that using the word "death" is prohibited by itself regardless of context, leading to idiotic phrasing in the style of "Person X was born on [date] and d!ed [other date]" that you see all over IG, FB and YT nowadays.)

    Apparently to reactivate the account I would need to prove that "[I am] a human being". That might be a tall order given that there are days when I doubt that myself.

    The reactivation service is designed in the usual deceptive way where it does not tell you all the things you need to do in advance. Instead it bounces you from one task to another in the hopes that sunk cost fallacy makes you submit to ever more egregious demands. I got out when they demanded a full video selfie where I look around different directions. You can make up your own theories as to why Meta, a known advocate for generative AI and all that garbage, would want a high resolution scans of people's faces. I mean, surely they would not use it for AI training without paying a single cent for usage rights to the original model. Right? Right?

    The suspension email ends with this ultimatum.

    If you think we suspended your account by mistake, you have 180 days to appeal our decision. If you miss this deadline your account will be permanently disabled.

    Well, mr Zuckerberg, my response is the following:

    Close it! Delete it! Burn it down to the ground! I'd do it myself this very moment, but I can't delete the account without reactivating it first.

    Let it also be noted that this post is a much better way of proving that I am a human being than a video selfie thing that could be trivially faked with genAI.

    • chevron_right

      Arun Raghavan: Accessibility Update: Enabling Mono Audio

      news.movim.eu / PlanetGnome • 2 days ago - 00:09 • 2 minutes

    If you maintain a Linux audio settings component, we now have a way to globally enable/disable mono audio for users who do not want stereo separation of their audio (for example, due to hearing loss in one ear). Read on for the details on how to do this.

    Background

    Most systems support stereo audio via their default speaker output or 3.5mm analog connector. These devices are exposed as stereo devices to applications, and applications typically render stereo content to these devices.

    Visual media use stereo for directional cues, and music is usually produced using stereo effects to separate instruments, or provide a specific experience.

    It is not uncommon for modern systems to provide a “mono audio” option that allows users to have all stereo content mixed together and played to both output channels. The most common scenario is hearing loss in one ear.

    PulseAudio and PipeWire have supported forcing mono audio on the system via configuration files for a while now. However, this is not easy to expose via user interfaces, and unfortunately remains a power-user feature.

    Implementation

    Recently, Julian Bouzas implemented a WirePlumber setting to force all hardware audio outputs (MR 721 and 769 ). This lets the system run in stereo mode, but configures the audioadapter around the device node to mix down the final audio to mono.

    This can be enabled using the WirePlumber settings via API, or using the command line with:

    wpctl settings node.features.audio.mono true

    The WirePlumber settings API allows you to query the current value as well as clear the setting and restoring to the default state.

    I have also added ( MR1 , MR2 ) a mechanism to set this using the PulseAudio API (via the messaging system). Assuming you are using pipewire-pulse , PipeWire’s PulseAudio emulation daemon, you can use pa_context_send_message_to_object() or the command line:

    pactl send-message /core pipewire-pulse:force-mono-output true

    This API allows for a few things:

    • Query existence of the feature: when an empty message body is sent, if a null value is returned, feature is not supported
    • Query current value: when an empty message body is sent, the current value ( true or false ) is returned if the feature is supported
    • Setting a value: the requested setting ( true or false ) can be sent as the message body
    • Clearing the current value: sending a message body of null clears the current setting and restores the default

    Looking ahead

    This feature will become available in the next release of PipeWire (both 1.4.10 and 1.6.0).

    I will be adding a toggle in Pavucontrol to expose this, and I hope that GNOME, KDE and other desktop environments will be able to pick this up before long.

    Hit me up if you have any questions!

    • chevron_right

      Allan Day: GNOME Foundation Update, 2026-01-09

      news.movim.eu / PlanetGnome • 6 days ago - 15:56 • 3 minutes

    Welcome to the first GNOME Foundation update of 2026! I hope that the new year finds you well. The following is a brief summary of what’s been happening in the Foundation this week.

    Trademark registration renewals

    This week we received news that GNOME’s trademark registration renewals have been completed. This is an example of the routine legal functions that the GNOME Foundation handles for the GNOME Project, and is part of what I think of as our core operations. The registration lasts for 10 years, so the next renewal is due in 2036. Many thanks to our trademark lawyers for handling this for us!

    Microsoft developer account

    Another slow registration process that completed this week was getting verified status on our Microsoft Developer Account. This was primarily being handled by Andy Holmes, with a bit of assistance on the Foundation side, so many thanks to him. The verification is required to allow those with Microsoft 365 organizational accounts to use GNOME Online Accounts.

    Travel Committee

    The Travel Committee had its first meeting of 2026 this week, where it discussed travel sponsorships for last month’s GNOME.Asia conference. Sadly, a number of people who were planning to travel to the conference had their visas denied. The committee spent some time assessing what happened with these visa applications, and discussed how to support visa applicants better in future. Thanks in particular to Maria for leading that conversation.

    GNOME.Asia Report

    Also related to GNOME.Asia: Kristi has posted a very nice report on the event , including some very nice pictures. It looks like it was a great event! Do make sure that you check out the post.

    Audit preparation

    As I mentioned in previous posts, audit preparation is going to be a major focus for the GNOME Foundation over the next three months. We are also finishing off the final details of our 2024-25 accounts. These two factors resulted in a lot of activity around the books this week. In addition to a lot of back and forth with our bookkeeper and finance advisor, we also had a regular monthly bookkeeping call yesterday, and will be having an extra meeting to make more process in the next few weeks.

    New payments platform rollout

    With it being the first week of the month, we had a batch of invoices to process and pay this week. For this we made the switch to a new payments processing system, which is going to be used for reimbursement and invoice tracking going forward. So far the system is working really well, and provides us with a more robust, compliant, and integrated process than what we had previously.

    Infrastructure

    Over the holiday, Bart cleared up the GNOME infrastructure issues backlog. This led him to write a service which will allow us to respond to GitLab abuse reports in a better fashion. On the Flathub side, he completed some work on build reproducibility, and finished adding the ability to re-publish apps that were previously marked as end of life.

    FOSDEM

    FOSDEM 2026 preparations continued this week. We will be having an Advisory Board meeting, for which attendance is looking good, so good that we are currently in the process of booking a bigger room. We are also in the process of securing a venue for a GNOME social event on the Saturday night.

    GNOME Foundation donation receipts

    Bart added a new feature to donate.gnome.org this week, to allow donors to generate a report on their donations over the last calendar year. This is intended to provide US tax payers with the documentation necessary to allow them to offset their donations against their tax payments. If you are a donor, you can generate a receipt for 2025 at donate.gnome.org/help .

    That’s it for this week’s update! Thanks for reading, and have a great weekend.

    • chevron_right

      Jussi Pakkanen: AI and money

      news.movim.eu / PlanetGnome • 6 days ago - 13:56 • 3 minutes

    If you ask people why they are using AI (or want other people to use it) you get a ton of different answers. Typically none of them contain the real reason, which is that using AI is dirt cheap. Between paying a fair amount to get something done and paying very little to give off an impression that the work has been done, the latter tends to win.

    The reason AI is so cheap is that it is being paid by investors. And the one thing we know for certain about those kinds of people is that they expect to get their money back. Multiple times over. This might get done by selling the system to a bigger fool before it collapses, but eventually someone will have to earn that money back from actual customers (or from government bailouts, i.e. tax payers).

    I'm not an economist and took a grand total of one economics class in the university, most of which I have forgotten. Still, using just that knowledge we can get a rough estimate of the money flows involved. For simplicity let's bundle all AI companies to a single entity and assume a business model based on flat monthly fees.

    The total investment

    A number that has been floated around is that AI companies have invested approximately one trillion (one thousand billion or 1e12) dollars. Let's use that as the base investment we want to recover.

    Number of customers

    Sticking with round figures, let's assume that AI usage becomes ubiquitous and that there are one billion monthly subscribers. For comparison the estimated number of current Netflix subscribers is 300 million.

    Income and expenses

    This one is really hard to estimate. What seems to be the case is that current monthly fees are not enough to even pay back the electricity costs of providing the service. But let's again be generous and assume that some sort of a efficiency breakthrough happens in the future and that the monthly fee is $20 with expenses being $10. This means a $10 profit per user per month.

    We ignore one-off costs such as buying several data centers' worth of GPUs every few years to replace the old ones.

    The simple computation

    With these figures you get $10 billion per month or $120 billion per year. Thus paying off the investment would take a bit more than 8 years. I don't personally know any venture capitalists, but based on random guessing this might fall in the "takes too long, but just about tolerable" level of delay.

    So all good then?

    Not so fast!

    One thing to keep in mind when doing investment payback calculations is the time value of money . Money you get in "the future" is not as valuable as money you have right now. Thus we need to discount them to current value.

    Interest rate

    I have no idea what a reasonable discount rate for this would be. So let's pick a round number of 5.

    The "real-er" numbers

    At this point the computations become complex enough that you need to break out the big guns. Yes, spreadsheets.

    Here we see that it actually takes 12 years to earn back the investment. Doubling the investment to two trillion would take 36 years. That is a fair bit of time for someone else to create a different system that performs maybe 70% as well but which costs a fraction of the old systems to get running and operate. By which time they can drive the price so low that established players can't even earn their operating expenses let alone pay back the original investment.

    Exercises for the reader

    • This computation assumes the system to have one billion subscribers from day one. How much longer does it take to recuperate the investment if it takes 5 years to reach that many subscribers? What about 10 years?
    • How long is the payback period if you have a mere 500 million paid subscribers?
    • Your boss is concerned about the long payback period and wants to shorten it by increasing the monthly fee. Estimate how many people would stop using the service and its effect on the payback time if the fee is raised from $20 to $50. How about $100? Or $1000?
    • What happens when the ad revenue you can obtain by dumping tons of AI slop on the Internet falls below the cost of producing said slop?
    • chevron_right

      Engagement Blog: GNOME ASIA 2025-Event Report

      news.movim.eu / PlanetGnome • 6 days ago - 11:35 • 1 minute

    GNOME ASIA 2025 took place in Tokyo, Japan, from 13–14 December 2025 , bringing together the GNOME community for the featured annual GNOME conference in Asia.
    The event was held in a hybrid format, welcoming both in-person and online speakers and attendees from across the world.

    GNOME ASIA 2025 was co-hosted with the LibreOffice Asia Conference community event , creating a shared space for collaboration and discussion between open-source communities.

    Photo by Tetsuji Koyama, licensed under CC BY 4.0

    About GNOME.Asia Summit

    The GNOME.Asia Summit focuses primarily on the GNOME desktop while also covering applications and platform development tools. It brings together users, developers, foundation leaders, governments, and businesses in Asia to discuss current technologies and future developments within the GNOME ecosystem.

    The event featured 25 speakers in total , delivering 17 full talks and 8 lightning talks across the two days. Speakers joined both on-site and remotely.

    Photo by Tetsuji Koyama, licensed under CC BY 4.0

    Around 100 participants attended in person in Tokyo, contributing to engaging discussions and community interaction. Session recordings were published on the GNOME Asia YouTube channel , where they have received 1,154 total views , extending the reach of the event beyond the conference dates.

    With strong in-person attendance, active online participation, and collaboration with the LibreOffice Asia community, GNOME ASIA 2025 once again demonstrated the importance of regional gatherings in strengthening the GNOME ecosystem and open-source collaboration in Asia.

    Photo by Tetsuji Koyama, licensed under CC BY 4.0

    • chevron_right

      This Week in GNOME: #231 Blueprint Maps

      news.movim.eu / PlanetGnome • 6 days ago - 00:00 • 4 minutes

    Update on what happened across the GNOME project in the week from January 02 to January 09.

    GNOME Core Apps and Libraries

    Maps

    Maps gives you quick access to maps all across the world.

    mlundblad announces

    Thanks to work done by Jamie Gravendeel Maps has now been ported to use Blueprint to define the UI templates. Also Hari Rana ported the share locations (“Send to”) dialog to AdwDialog.

    send_to_dialog_adwaita.BTYip87X_Z1BKLEf.webp

    send_to_dialog_adwaita_mobile.DC_To8Oz_1x7pG2.webp

    Third Party Projects

    Giant Pink Robots! says

    Version v2026.1.5 of the Varia download manager was released with automatic archive extraction, improvements to accessibility and tons of bug fixes and small improvements. The biggest part of this new release however is macOS support, albeit in an experimental state for now. With this, Varia now supports all three big desktop OS platforms: Linux, Windows and Mac. https://giantpinkrobots.github.io/varia/

    francescocaracciolo announces

    Newelle, AI Assistant for Gnome, received a new major update!

    • Added MCP server support, enabling integration with thousands of apps
    • Added Tools, extensions can now add new tools very easily
    • Added the possibility to set some models as favoutites
    • You can now trigger recording and TTS stop with keyboard shortcuts

    Download it on Flathub

    Phosh

    A pure wayland shell for mobile devices.

    Guido announces

    Phosh 0.52 is out:

    We’ve added a QR code to the Wi-Fi quick setting so clients can connect easily by scanning it and there’s a new gesture to control brightness on the lock screen.

    There’s more — see the full details here .

    hotspot-qrcode.gmTLfhOG_sxF8u.webp

    Flare

    Chat with your friends on Signal.

    schmiddi announces

    Version 0.18.0-beta.1 of Flare was now released on flathub-beta. This release includes fixes for using Flare as a primary device, which I have done successfully for a while now. Feel free to test it out and provide feedback. Note that if you want to try it out, I would heavily encourage linking Signal-Desktop to Flare in order to set your profile information and to start new chats. Feel free to give feedback if you have any issues with this beta in the Matrix room or issue tracker .

    Emergency Alerts

    Receive emergency alerts

    Leonhard reports

    Emergency Alerts 2.0.0 has been released! It finally brings the long-awaited weather alerts for the U.S. and air raid alerts for Ukraine. Location selection is now also more powerful, allowing you to choose any point on Earth, and the new map view lets you see active alerts and affected areas at a glance. Please note that to make all this possible, the way locations are stored had to be updated. When you first launch the app after updating, it tries to migrate your existing locations automatically. In rare cases, this may not work and you might need to re-add them manually. If that happens a notification will be sent.

    Highlights:

    • Weather alerts now available across the U.S.
    • Air raid alerts now available for Ukraine
    • Pick any point on Earth as a location
    • New map view showing active alerts and impacted areas

    emergency_alerts_details_about_an_alert.CujjRlPN_Z2aN3lQ.webp

    emergency_alerts_the_dashboard_with_locations_supported_by_the_new_providers.CVay8juq_9xA8C.webp

    emergency_alerts_the_new_map_view.CNFhyyqI_Z1Qel0A.webp

    GNOME Websites

    Sophie (she/her) says

    The www.gnome.org pages are now available in English, Bulgarian, Basque, Brazilian Portuguese, Swedish, Ukrainian, and Chinese. You can contribute additional translations on l10n.gnome.org .

    www.gnome.org.BLfru-JW_ZBo6vS.webp

    Miscellaneous

    Guillaume Bernard reports

    Damned Lies has been refreshed during the last weeks of 2025.

    To refresh the statistics of branches, many of you complained that the task was synchronous and ended in timeouts. I have reworked this part in anticipation of ticket #409 (asynchronous git pushes) and the refresh now delegates refresh statistics to a Celery worker. For git pushes, we’ll use Celery tasks the same way!

    In short, this means every time you click the refresh statistics button, it will start a job in the background, and a progress bar will show you the refresh status of the job in real time. There will be a maximum of three concurrent refreshes at a time, that should be enough :-).

    In addition to these major changes, I reworked the presentation of languages and POT files in modules:

    1. The date & time of the POT file generation is now shown with the number of messages.

    2. Your languages are shown on top of the list; it will no longer be necessary to scroll down to find your language in the language list.

    translation_2.7ZnJHQVY_4knME.webp

    translation.D419_v0W_ZUSVh1.webp

    Arjan reports

    PyGObject 3.55.1 has been released. It’s the second development release (it’s not available on PyPI) in the current GNOME release cycle.

    Notable changes include:

    • A fix do do_dispose() is always called on your object.
    • You can define a do_constructed() method that will be called after the object is initialised.
    • A regression in 3.55.0 has been fixed: instance data is now saved and outlives the garbage collector.

    All changes can be found in the Changelog

    This release can be downloaded from Gitlab and the GNOME download server .If you use PyGObject in your project, please give it a swing and see if everything works as expected.

    That’s all for this week!

    See you next week, and be sure to stop by #thisweek:gnome.org with updates on your own projects!

    • chevron_right

      Daiki Ueno: GNOME.Asia Summit 2025

      news.movim.eu / PlanetGnome • 7 January • 2 minutes

    Last month, I attended the GNOME.Asia Summit 2025 held at the IIJ office in Tokyo. This was my fourth time attending the summit, following previous events in Taipei (2010), Beijing (2015), and Delhi (2016).

    As I live near Tokyo, this year’s conference was a unique experience for me: an opportunity to welcome the international GNOME community to my home city rather than traveling abroad. Reconnecting with the community after several years provided a helpful perspective on how our ecosystem has evolved.

    Addressing the post-quantum transition

    During the summit, I delivered a keynote address regarding post-quantum cryptography (PQC) and desktop. The core of my presentation focused on the “Harvest Now, Decrypt Later” (HNDL) type of threats, where encrypted data is collected today with the intent of decrypting it once quantum computing matures. The talk was followed by the history and the current status of PQC support in crypto libraries including OpenSSL, GnuTLS, and NSS, and concluded with the next steps recommended for the users and developers.

    It is important to recognize that classical public key cryptography, which is vulnerable to quantum attacks, plays an integral role on the modern desktop: from secure web browsing to the underlying verification of system updates. Given that major government timelines (such as NIST and the NSA’s CNSA 2.0 ) are pushing for a full migration to quantum-resistant algorithms between 2027 and 2035, the GNU/Linux desktop should prioritize “crypto-agility” to remain secure in the coming decade.

    From discussion to implementation: Crypto Usage Analyzer

    One of the tools I discussed during my talk was crypto-auditing , a project designed to help developers identify and update the legacy cryptography usage. At the time of the summit, the tool was limited to a command-line interface, which I noted was a barrier to wider adoption.

    Inspired by the energy of the summit, I spent part of the recent holiday break developing a GUI for crypto-auditing. By utilizing AI-assisted development tools, I was able to rapidly prototype an application, which I call “Crypto Usage Analyzer” , that makes the auditing data more accessible.

    Conclusion

    The summit in Tokyo had a relatively small audience, which resulted in a cozy and professional atmosphere. This smaller scale proved beneficial for technical exchange, as it allowed for focused discussions on desktop-related topics than is often possible at larger conferences.

    Attending GNOME.Asia 2025 was a reminder of the steady work required to keep the desktop secure and relevant. I appreciate the efforts of the organizing committee in bringing the summit to Tokyo, and I look forward to continuing my work on making security libraries and tools more accessible for our users and developers.