Feeds
MidCamp - Midwest Drupal Camp: Last Chance Proposal Help: MidCamp 2025 Session Proposal Workshop
Missed the last Session Proposal Workshop? Don't worry; we have another one in January right before the submission deadline!
đ Ready to take your session ideas to the next level? Whether you're a seasoned speaker or a first-time presenter, the MidCamp 2025 Session Proposal Workshop is here to help you craft standout submissions.
đ
Date: January 7, 2025
đ Time: 3:00 PM - 4:00 PM CST
đ Location: Virtual via MidCamp Slack (#speakers channel)
This workshop will be led by Aaron Feledy, a seasoned Drupal contributor and experienced speaker. Aaron brings years of expertise in proposal crafting and conference speaking, offering practical advice to help you refine and elevate your session submissions.
Why Attend?Submitting a session proposal can be dauntingâbut it doesn't have to be! This workshop is designed to guide you through the process, from brainstorming topics to refining your submission. Our expert facilitators will share insider tips on what makes a proposal stand out to reviewers and resonate with attendees.
What Youâll Learn:- How to choose and frame a compelling topic
- Crafting clear, concise, and engaging abstracts
- Tips for tailoring your proposal to different audiences
- Insight into the MidCamp review process
Ready to submit? Session submissions for MidCamp 2025 are now open! Visit the MidCamp 2025 session submission page for guidelines and start your journey to the stage.
How to Join:Simply join the MidCamp Slack and head over to the #speakers channel on December 12th at 3:00 PM CST. No registration requiredâjust jump in and start collaborating!
Old New Blog
I started this blog back in 2010. Back then I used Wordpress and it worked reasonably well. In 2018 I decided to switch to a static generated site, mostly because the Wordpress blog felt slow to load and it was hassle to maintain. Back then the go-to static site generator was Jekyll, so I went with that. Lately I’ve been struggling with it though, because in order to keep all the plugins working, I needed to use older versions or Ruby, which meant I had to use Docker to build the blog locally. Overall, it felt like too much work and for the past few years I’ve been eyeing Hugo - more so since Carl and others migrated most of KDE websites to it. I mean, if it’s good enough for KDE, it’s good enough for me, right?
So this year I finally got around to do the switch. I migrated all the content from Jekyll. This time I actually went through every single post, converted it to proper Markdown, fixed formatting, images etc. It was a nice trip down the memory lane, reading all the old posts, remembering all the sprints and Akademies… I also took the opportunity to clean up the tags and categories, so that they are more consistent and useful.
Finally, I have rewritten the theme - I originally ported the template from Wordpress to Jekyll, but it was a bit of a mess, responsivity was “hacked” in via JavaScript. Web development (and my skills) has come a long way since then, so I was able to leverage more modern CSS and HTML features to make the site look the same, but be more responsive and accessible.
CommentsWhen I switched from Wordpress to Jekyll, I was looking for a way to preserve comments. I found Isso, which is basically a small CGI server backed with SQLite that you can run on the server and embed it into your static website through JavaScript. It could also natively import comments from Wordpress, so that’s the main reason why I went with it, I think. Isso was not perfect (although the development has picked up again in the past few years) and it kept breaking for me. I think it haven’t worked for the past few years on my blog and I just couldn’t be bothered to fix it. So, I decided to ditch it in favor of another solution…
I wanted to keep the comments for old posts by generating them as static HTML from the Isso’s SQLite database, alas the database file was empty. Looks like I lost all comments at some point in 2022. It sucks, but I guess it’s not the end of the world. Due to the nature of how Isso worked, not even the Wayback Machine was able to archive the comments, so I guess they are lost forever…
For this new blog, I decided to use Carl’s approach with embedding replies to a Mastodon. I think it’s a neat idea and it’s probably the most reliable solution for comments on a static blog (that I don’t have to pay for, host myself or deal with privacy concerns or advertising).
I have some more ideas regarding the comments system, but that’s for another post ;-) Hopefully I’ll get to blog more often now that I have a shiny new blog!
Happy Holidays đEnjoy the holidays and see you in 2025 đ„ł!
ImageX: Exploring New Features for Content Editors in Drupal as We Step into 2025
Freelock Blog: Automatically post to BlueSky
Since the 2024 election, the BlueSky social network has exploded in popularity, and appears to be replacing the cesspool that used to be Twitter. I'm not much of a social media person -- I much prefer hanging out in smaller spaces with people with shared interests. If you're like me, I would highly recommend finding a Mastodon server that caters to your interests, where you're sure to find rewarding conversations.
Noah Meyerhans: Local Development VM Management
A coworker asked recently about how people use VMs locally for dev work, so I figured Iâd take a few minutes to write up a bit about what I do. There are many use cases for local virtual machines in software development and testing. Theyâre self-contained, meaning you can make a mess of them without impacting your day-to-day computing environment. They can run different distributions, kernels, and even entirely different operating systems from the one you use regularly. Etc. Theyâre also cheaper than cloud services and provide finer grained control over the resources.
I figured Iâd share a little bit about how I manage different virtual machines in case anybody finds this useful. This is what works for me, but it wonât necessarily work for you, or maybe youâve already got something better. Iâve found it to be easy to work with, light weight, and is easy to evolve my needs change.
Use short-lived VMsRather than keep a long-lived âdevelopmentâ VM around that you customize over time, I recommend automating the common customizations and provisioning new VMs regularly. If Iâm working on reproducing a bug or testing a change prior to submitting it upstream, Iâll do this work in a VM and delete the VM when when Iâm done. When provisioning VMs this frequently, though, walking through the installation process for every new VM is tedious and a waste of time. Since most of my work is done in Debian, so I start with images generated daily by the cloud team. These images are available for multiple releases and architectures. The ânocloudâ variant boots to a root prompt and can be useful directly, or the âgenericâ images can be used for cloud-init based customization.
Automating image preparationThis makefile lets me do something like make image and get a new qcow2 image with the latest build of a given Debian release (sid by default, with others available by specifying DIST).
DATESTAMP=$(shell date +"%Y-%m-%d") FLAVOR?=generic ARCH?=$(shell dpkg --print-architecture) DIST?=sid RELEASE=$(DIST) URL_PATH=https://cloud.debian.org/images/cloud/$(DIST)/daily/latest/ ifeq ($(DIST),trixie) RELEASE=13 endif ifeq ($(DIST),bookworm) RELEASE=12 endif ifeq ($(DIST),bullseye) RELEASE=11 endif debian-$(DIST)-$(FLAVOR)-$(ARCH)-daily.tar.xz: curl --fail --connect-timeout 20 -LO \ $(URL_PATH)/debian-$(RELEASE)-$(FLAVOR)-$(ARCH)-daily.tar.xz $(DIST)-$(FLAVOR)-$(DATESTAMP).qcow2: debian-$(RELEASE)-$(FLAVOR)-$(ARCH)-daily.tar.xz tar xvf debian-$(RELEASE)-$(FLAVOR)-$(ARCH)-daily.tar.xz qemu-img convert -O qcow2 disk.raw $@ rm -f disk.raw qemu-img resize $@ 20g qemu-img snapshot -c untouched $@ image: $(DIST)-$(FLAVOR)-$(DATESTAMP).qcow2 .PHONY: image Customize the VM environment with cloud-initWhile the ânocloudâ images can be useful, I typically find that I want to apply the same modifications to each new VM I launch, and they donât provide facilities for automating this. The âgenericâ images, on the other hand, run cloud-init by default. Using cloud-init, I can create my user account, point apt at local mirrors, install my preferred tools, ensure the root filesystem is resized to make full use of the backing storage, etc.
The cloud-init configuration on the generic images will read from a local config drive, which can contain an ISO9660 (cdrom) filesystem image. This image can be generated from a subdirectory containing the various cloud-init input files using the following make syntax:
IMDS_FILES=$(shell find seedconfig -path '*/.git/*' \ -prune -o -type f -name '*.in.json' -print) \ seedconfig/openstack/latest/user_data seed.iso: $(IMDS_FILES) genisoimage -V config-2 -o $@ -J -R -m '*~' -m '.git' seedconfigWith the image in place, the VM can be created with
qemu-system-x86_64 -machine q35,accel=kvm -cpu host -m 4g -drive file=${img},index=0,if=virtio,media=disk -drive file=seed.iso,media=cdrom,format=raw,index=2,if=virtio -nic user -nographicThis invokes qemu with the root volume and ISO image attached as disks, uses an emulated âq35â machine with the hostâs CPU and KVM acceleration, the userspace network stack, and a serial console. The first time the VM boots, cloud-init will apply the configuration from the cloud-config available in the ISO9660 filesystem.
Alternatives to cloud-initvirt-customize is another tool accomplishing the same type of customization. I use cloud-init because it works directly with cloud providers in addition to local VM images. You could also use something like ansible.
VariationsI have a variant of this that uses a bridged network, which Iâll write more about later. The bridge is nice because itâs more featureful, with full support for IPv6, etc, but it needs a bit more infrastructure in place.
It also can be helpful to use 9p or virtfs to share filesystem state between the host the VM. I donât tend to rely on these, and will instead use rsync or TRAMP for moving files around.
Containers are also useful, of course, and there are plenty of times when the full isolation of a VM is not worth the overhead.
The Drop Times: An Enriching Experience to Carry Forward: Reflections from DrupalCon Asia
Web Review, Week 2024-51
Let’s go for my web review for the week 2024-51.
Advice for First-Time Open Source ContributorsTags: tech, foss, community
Definitely a good list of advices for first time contributors.
https://www.yegor256.com/2024/12/15/open-source-beginner-advice.html
Tags: tech, internet, geospatial
IRISÂČ is the friendly reminder that tens of thousand of low orbit satellites is not the only design… and likely not the smartest one.
https://www.theverge.com/2024/12/16/24322358/iris2-starlink-rival-europe-date-cost
Tags: tech, tv, attention-economy, advertisement
The TV market is really turning into an anti-consumer one.
Tags: tech, social-media, bluesky, fediverse, architecture
Yet another long piece in this interesting and in depth conversation about Bluesky. The fact that it stays civil is called out explicitly and this is appreciated.
https://dustycloud.org/blog/re-re-bluesky-decentralization/
Tags: tech, social-media, moderation, bluesky, politics
Bluesky is already hitting growth pains regarding moderation and its guidelines. By being centralized it is also more at risk within the current US political climate.
Tags: tech, social-media, linkedin, ai, machine-learning, gpt, fake
Kind of unsurprising right? I mean LinkedIn is clearly a deformed version of reality where people write like corporate drones most of the time. It was only a matter of time until robot generated content would be prevalent there, it’s just harder to spot since even humans aren’t behaving genuinely there.
https://www.wired.com/story/linkedin-ai-generated-influencers/
Tags: tech, internet, web, ai, machine-learning, gpt, fake, knowledge
Indeed, we’ll have to relearn “internet hygiene”, it is changing quickly now that we prematurely unleashed LLM content on the open web.
https://www.late-review.com/p/ai-and-internet-hygiene
Tags: tech, ai, machine-learning, gpt, criticism
A good balanced post on the topic. Maybe we’ll finally see a resurgence of real research innovation and not just stupid scaling at all costs. Reliability will stay the important factor of course and this one is still hard to crack.
https://www.aisnakeoil.com/p/is-ai-progress-slowing-down
Tags: tech, analogic, ai, machine-learning, neural-networks, hardware
It looks like analog chips for neural network workloads are on the verge of finally becoming reality. This would reduce consumption by an order of magnitude and hopefully more later on. Very early days for this new attempt, let’s see if it holds its promises.
https://spectrum.ieee.org/analog-ai-2669898661
Tags: tech, hardware, foss
A good question, it is somewhat of a grey area at times. We need to come up with better answers.
https://mjg59.dreamwidth.org/70895.html
Tags: tech, databases, sqlite, asynchronous, rust, system, filesystem
Interesting explanation of a research paper exploring the possibility of a faster SQLite by focusing on async I/O.
https://avi.im/blag/2024/faster-sqlite/
Tags: tech, java, tools
I wouldn’t use it as much as advocated in this article, still this is a good reminder that Java became way more approachable for smaller programs in recent years.
https://horstmann.com/unblog/2024-12-11/index.html
Tags: tech, career, complexity, learning
It tries hard at not being a “get off my lawn” post. It clearly points some kind of disconnects though. They’re real. I guess it’s to be expected with the breadth of our industry. There are so many abstractions piled onto each other that it’s difficult to explore them all.
https://rakhim.exotext.com/web-developers-a-growing-disconnect
Tags: tech, design, pattern, java, type-systems
One of my favorite of the traditional design patterns in object oriented languages. Now obviously when you get pattern matching in your language… you don’t need the visitor pattern anymore.
https://nipafx.dev/java-visitor-pattern-pointless/
Tags: tech, project-management, estimates
I don’t exactly use this approach to factor in the uncertainty… but I guess there’s something to be made out of this proposal. I’ll keep it in mind for my next project.
https://ntietz.com/blog/estimating-projects-short-sale/
Tags: leadership, management, communication
Interesting ideas about leadership lacking in impact. Indeed it should be seen as a communal function, it’s not about individuals leading each in their own directions. Think about it in a systemic way.
https://suzansfieldnotes.substack.com/p/the-one-way-i-know-a-team-is-in-trouble
Tags: ecology, politics, law, energy
This is not all bad news, there are a few things to rejoice about.
Bye for now!
Real Python: The Real Python Podcast â Episode #232: Exploring Modern Sentiment Analysis Approaches in Python
What are the current approaches for analyzing emotions within a piece of text? Which tools and Python packages should you use for sentiment analysis? This week, Jodie Burchell, developer advocate for data science at JetBrains, returns to the show to discuss modern sentiment analysis in Python.
[ Improve Your Python With đ Python Tricks đ â Get a short & sweet Python Trick delivered to your inbox every couple of days. >> Click here to learn more and see examples ]
LostCarPark Drupal Blog: Drupal Advent Calendar day 20 - Navigation
Itâs day 20 of the Drupal Advent Calendar, and today weâre looking at the admin UI Navigation. Joining us today are Pablo LĂłpez and Matthew Oliveira, so letâs look into itâŠ
The aim of the Navigation track is to provide a better site management experience for Drupal users. It does not provide a specific recipe or feature to Drupal CMS. Navigation is a core experimental module. However, the Navigation track provides key integration points to Drupal CMS that will help other tracks to highlight their features in the new Navigation left sidebar.
The navigation sidebar provides an improved interface for site builders and content creatorsSince Navigation has replaced Toolbar in Drupal CMSâŠ
TagsCKEditor: Unlock New Levels of Drupal Content Editing: Webinar Recap
Talk Python to Me: #489: Anaconda Toolbox for Excel and more with Peter Wang
New LabPlot User Documentation
In recent weeks we have been working on transferring LabPlot’s documentation to a new format.
We decided to move the documentation from the DocBook and MediaWiki format to the Sphinx/reStrcutredText framework. In our perception Sphinx offers a user-friendly and flexible way to create and manage documentation. Easy math typing and code formatting also come along. Additionally, Sphinx supports basic syntax checks, and modern documentation practices, such as versioning and integration with various output formats like HTML, PDF and ePub.
The new user’s manual is available on a dedicated page: https://docs.labplot.org. Please check it out and let us know what you think.
The manual still needs to be supplemented with new content, so we encourage you to contribute to the documentation, e.g. by fixing and adding new sections, updating images, as collaborative efforts can lead to a more comprehensive resource for everyone. Please check the Git repository dedicated to the documentation to find more details on how to help make it better.
Matt Layman: Bootstrap Kamal On Droplet - Building SaaS #209.1
Matt Layman: Docker Image For Kamal - Building SaaS #209.2
Matt Layman: Postgres To SQLite - Building SaaS #209.3
Digging into the Fast Sketch Cleanup Plugin for Krita
We started this project with the intent of providing users a tool helpful in inking sketches. It is based on a research article by Simo & Sierra published in 2016, and it uses neural networks (now commonly called simply AI) to work. The tool has been developed in partnership with Intel and itâs still considered experimental, but you can already use it and see the results.
In the section below there are some real life examples of use cases and the results from the plugin. The results vary, but it can be used for extracting faint pencil sketches from photos, cleaning up lines, and comic book inking.
Regarding the model used in the tool, we trained it ourselves. All the data in the dataset is donated from people who sent their pictures to us themselves and agreed on this specific use case. We havenât used any other data. Moreover, when you use the plugin, it processes locally on your machine, it doesnât require any internet connection, doesnât connect to any server, and no account is required either. Currently it works only on Windows and Linux, but weâll work on making it available on MacOS as well.
Use casesIt averages the lines into one line and creates strong black lines, but the end result can be blurry or uneven. In many cases however it still works better than just using a Levels filter (for example in extracting the pencil sketch). it might be a good idea to use Levels filter after using the plugin to reduce the blurriness. Since the plugin works best with white canvas and grey-black lines, in case of photographed pencil sketches or very light sketch lines, it might be a good idea to use Levels also before using the plugin.
Extracting photographed pencil sketchThis is the result of the standard procedure of using Levels filter on a sketch to extract the lines (which results in a part of the image getting the shadow):
sketch_girl_original_procedure_comparison_small1843Ă1209 165 KB
The sketch was drawn by Tiar (link to KA profile)
This is the procedure using the plugin with SketchyModel (Levels â plugin â Levels):
sketch_girl_new_procedure_comparison_small1843Ă2419 267 KB
Comparison (for black lines):
sketch_girl_procedures_comparison_small1920Ă1260 215 KB
Another possible result is to just stop at the plugin without forcing black lines using Levels, which results in a nicer, more pencil-y look while keeping the lower part of the page still blank:
sketch_girl_after_plugin_small1536Ă2016 161 KB
Picture of a man made by BeARToys
Here in the pictures above you can see the comic book style inking. The result, which is a bit blurry compared to the original, can be further enhanced by using a Sharpen filter. The dragon was sketched by David Revoy (CC-BY 4.0).
Cleaning up linesExamples of sketches I made and the result of the plugin, showing the strong and weak points of the plugin. All of the pictures below were made using the SketchyModel.
flower_001_detail681Ă456 22.1 KB
portrait_man_portrait_2_comparison_2_small1305Ă505 139 KB
portrait_man_portrait_2_detail646Ă1023 26.6 KB
All of the pictures above painted by Tiar (link to KA profile)
On the pictures below, on the scales of the fish, you can see how the model discriminates lighter lines and enhances the stronger lines, making the scales more pronounced. In theory you could do that using the Levels filter, but in practice the results would be worse, because the model takes into account local strength of the line.
fish_square_sketchy_comparison_small1920Ă968 156 KB
Picture of the fish made by Christine Garner (link to portfolio)
How to use it in KritaTo use the Fast Sketch Cleanup plugin in Krita, do the following:
-
Prepare Krita:
-
On Windows:
-
Either in one package: download Krita 5.3.0-prealpha with Fast Sketch Cleanup plugin already included: https://download.kde.org/unstable/krita/5.3.0-prealpha-fast-sketch/krita-x64-5.3.0-prealpha-cdac9c31.zip
-
Or separately:
- Download portable version of Krita 5.2.6 (or similar version - should still work)
- Download separately the Fast Sketch Cleanup plugin here: https://download.kde.org/stable/krita/FastSketchPlugin-1.0.2/FastSketchPlugin1.0.2.zip
- Unzip the file into krita-5.2.6/ folder (keeping the folder structure).
- Then go to Settings â Configure Krita â Python Plugin Manager, enable Fast Sketch Cleanup plugin, and restart Krita.
-
-
On Linux:
- Download the appimage: https://download.kde.org/unstable/krita/5.3.0-prealpha-fast-sketch/krita-5.3.0-prealpha-cdac9c31c9-x86_64.AppImage
-
-
(Optional) Install NPU drivers if you have NPU on your device (practically only necessary on Linux, if you have a very new Intel CPU): Configurations for IntelÂź NPU with OpenVINOâą â OpenVINOâą documentation (note: you can still run the plugin on CPU or GPU, it doesnât require NPU)
-
Run the plugin:
- Open or create a white canvas with grey-white strokes (note that the plugin will take the current projection of the canvas, not the current layer).
- Go to Tools â Fast Sketch Cleanup
- Select the model. Advanced Options will be automatically selected for you.
- Wait until it finishes processing (the dialog will close automatically then).
- See that it created a new layer with the result.
Currently itâs better to just use the SketchyModel.xml, in most cases it works significantly better than the SmoothModel.xml.
You need to make sure the background is pretty bright, and the lines you want to keep in the result are relatively dark (either somewhat dark grey or black; light grey might result in many missed lines). It might be a good idea to use a filter like Levels beforehand.
After processing, you might want to enhance the results with either Levels filter or Sharpen filter, depending on your results.
Technology & Science behind it Unique requirementsFirst unique requirement was that it had to work on canvases of all sizes. That meant that the network couldnât have any dense/fully or densely connected linear layers that are very common in most of the image processing neural networks (which require input of a specific size and will produce different results for the same pixel depending on its location), only convolutions or pooling or similar layers that were producing the same results for every pixel of the canvas, no matter the location. Fortunately, the Simo & Sierra paper published in 2016 described a network just like that.
Another challenge was that we couldnât really use the model they created, since it wasnât compatible with Kritaâs license, and we couldnât even really use the exact model type they described, because one of those model files would be nearly as big as Krita, and the training would take a really long time. We needed something that would work just as well if not better, but small enough that it can be added to Krita without making it twice as big. (In theory, we could do like some other companies and make the processing happen on some kind of a server, but that wasnât what we wanted. And even if it resolved some of our issues, it would provide plenty of its own major challenges. Also, we wanted for our users to be able to use it locally without a reliance on our servers and the internet). Moreover, the model had to be reasonably fast and also modest in regards to RAM/VRAM consumption.
Moreover, we didnât have any dataset we could use. Simo & Sierra used a dataset, where the expected images were all drawn using a constant line width and transparency, which meant that the results of the training had those qualities too. We wanted something that looked a bit more hand-drawn, with varying line-width or semi-transparent ends of the lines, so our dataset had to contain those kinds of images. Since we havenât been aware of any datasets that would match our requirements regarding the license and the data gathering process, we asked our own community for help, here you can read the Krita Artists thread about it: https://krita-artists.org/t/call-for-donation-of-artworks-for-the-fast-line-art-project/96401 .
The link to our full dataset can be found below in the Dataset section.
Model architectureAll main layers are either convolutional or deconvolutional (at the end of the model). After every (de)convolutional layer except for the last one there is a ReLu activation layer, and after the last convolution there is a sigmoid activation layer.
Python packages used: Pillow, Numpy, PyTorch and OpenvinoNumpy is a standard library for all kinds of arrays and advanced array operations and we used Pillow for reading images and converting them into numpy arrays and back. For training, we used PyTorch, while in the Krita plugin we used Openvino for inference (processing through the network).
Using NPU for inference
This table shows the result of benchmark_app, which is a tool thatâs provided with Intelâs python package openvino. It tests the model in isolation on random data. As you can see, the NPU was several times faster than the CPU on the same machine.
On the other hand, introducing NPU added a challenge: the only models that can run on NPU are static models, meaning the input size is known at the time of saving the model to file. To solve this, the plugin first cuts the canvas into smaller parts of a specified size (which depends on the model file), and then proceeds to process all of them and finally stitch the results together. To avoid artifacts on the areas next to the stitching, all of the parts are cut with a little bit of a margin and the margin is later cut off.
How to train your own modelTo train your own model, youâll need some technical skills, pairs of pictures (input and the expected output) and a powerful computer. You might also need quite a lot of space on your hard drive, though you can just remove unnecessary older models if you start having issues with lack of space.
Drivers & preparationYouâll need to install Python3 and the following packages: Pillow, openvino, numpy, torch. For quantization of the model you will also need nncf and sklearn. If I missed anything, it will complain, so just install those packages it mentions too.
If youâre on Windows, you probably have drivers for NPU and dedicated GPU. On Linux, you might need to install NPU drivers before youâll be able to use it: https://docs.openvino.ai/2024/get-started/configurations/configurations-intel-npu.html .
Moreover if you want to use iGPU for training (which might still be significantly faster than on CPU), youâll probably need to use something like IPEX which allows PyTorch to use an âXPUâ device, which is just your iGPU. Itâs not tested or recommended since I personally havenât been able to use it because my Python version was higher than the instruction expects, but the instruction is here: https://pytorch-extension.intel.com/installation?platform=gpu&version=v2.5.10%2Bxpu .
The sanity check for the installation is as follows:
python3 -c "import torch; import intel_extension_for_pytorch as ipex; print(f'Packages versions:'); print(f'Torch version: {torch.__version__}'); print(f'IPEX version: {ipex.__version__}'); print(f'Devices:'); print(f'Torch XPU device count: {torch.xpu.device_count()}'); [print(f'[Device {i}]: {torch.xpu.get_device_properties(i)}') for i in range(torch.xpu.device_count())];"
It should show more than 0 devices with some basic properties.
If you manage to get XPU device working on your machine, youâll still need to edit the training scripts so theyâll able to use it: https://intel.github.io/intel-extension-for-pytorch/xpu/latest/tutorials/getting_started.html (most probably youâll just need to add this line:
import intel_extension_for_pytorch as ipex
to the script on the very top, just underneath âimport torchâ, and use âxpuâ as the device name when invoking the script, and it should work. But as I said, the scripts hasnât been tested for that.
Youâll need some pictures to be able to train your model. The pictures must be in pairs, every pair must contain a sketch (input) and a lineart picture (expected output). The better quality of the dataset, the better the results.
Before training, itâs best if you augment the data: that means the pictures are rotated, scaled up or down, and mirrored. Currently the data augmentation script also performs an inversion with the assumption that training on inverted pictures would bring the results faster (considering that black means zero means no signal, and weâd like that to be the background, so the models learn the lines, not the background around lines).
How to use the data augmentation script is explained below in the detailed instruction for the training part.
Hereâs the dataset that we used (please read the license carefully if you want to use it): https://files.kde.org/krita/extras/FastSketchCleanupPluginKritaDataset.zip
Choice of model and other parametersFor quick results, use tooSmallConv; if you have more time and resources, typicalDeep might be a better idea. If you have access to a powerful GPU machine, you might try original or originalSmaller, which represent the original description of the model from the SIGGRAPH article by Simo-Sierra 2016, and a smaller version of it.
Use adadelta as the optimizer.
You can use either blackWhite or mse as the loss function; mse is classic, but blackWhite might lead to faster results since it lowers the relative error on the fully white or fully black areas (based on the expected output picture).
Training-
Clone the repository at https://invent.kde.org/tymond/fast-line-art (at 33869b6)
git clone https://invent.kde.org/tymond/fast-line-art.git -
Then, prepare the folder:
- Create a new folder for the training.
- In the folder, run:
python3 [repository folder]/spawnExperiment.py --path [path to new folder, either relative or absolute] --note "[your personal note about the experiment]"
-
Prepare data:
- If you have existing augmented dataset, put it all in data/training/ and data/verify/, keeping in mind that paired pictures in ink/ and sketch/ subfolders must have the exact same names (for example if you have sketch.png and ink.png as data, you need to put one in sketch/ as picture.png and another in ink/ as picture.png to be paired).
- If you don't have existing augmented dataset:
- Put all your raw data in data/raw/, keeping in mind that paired pictures should have the exact same names with added prefix either ink_ or sketch_ (for example if you have picture_1.png being the sketch picture and picture_2.png being the ink picture, you need to name them sketch_picture.png and ink_picture.png respectively.)
- Run the data preparer script:
python3 [repository folder]/dataPreparer.py -t taskfile.yml
That will augment the data in the raw directory in order for the training to be more successful.
-
Edit the taskfile.yml file to your liking. The most important parts you want to change are:
- model type - code name for the model type, use tinyTinier, tooSmallConv, typicalDeep or tinyNarrowerShallow
- optimizer - type of optimizer, use adadelta or sgd
- learning rate - learning rate for sgd if in use
- loss function - code name for loss function, use mse for mean squared error or blackWhite for a custom loss function based on mse, but a bit smaller for pixels where the target image pixel value is close to 0.5
-
Run the training code:
python3 [repository folder]/train.py -t taskfile.yml -d "cpu"On Linux, if you want it to run in a background, add â&â at the end. If it runs in a foreground, you can pause the training just by pressing ctrl+C, and if it runs in a background, find a process id (using either âjobs -lâ command or âps aux | grep train.pyâ command, the first number would be the process id) and kill it using âkill [process id]â command. Your results will still be in the folder, and youâll be able to resume the training using the same command.
-
Convert the model to an openvino model:
python3 [repository folder]/modelConverter.py -s [size of the input, recommended 256] -t [input model name, from pytorch] -o [openvino model name, must end with .xml] -
Place both the .xml and .bin model files in your Krita resource folder (inside pykrita/fast_sketch_cleanup subfolder) alongside other models to use them in the plugin.
Python Morsels: Merging dictionaries in Python
Merging dictionaries in Python is usually as simple as a single pipe character.
Table of contents
- Merging dictionaries with | in Python
- Merging dictionaries with the update method
- Merging dictionaries with **
- The difference between | and **
- Handling duplicate keys when merging
- The | operator performs a "union" between dictionaries
- Join dictionaries with the | operator
First let's talk about the simplest way to merge dictionaries, which will usually be all that you need.
Here are two dictionaries:
>>> context = {"language": "en", "timezone": "UTC"} >>> more_context = {"title": "Home", "breadcrumbs": ["Home"]}We'd like to make a new third dictionary that combines these two. This new dictionary should contain all the key-value pairs from both dictionaries.
The easiest way to do this is to use the pipe (|) operator:
>>> new_context = context | more_contextThis made a new dictionary that contains all the items from both of the two initial dictionaries:
>>> new_context {'language': 'en', 'timezone': 'UTC', 'title': 'Home', 'breadcrumbs': ['Home']}Using the | operator is basically the same as making a new empty dictionary, and then looping over all the items in the first dictionary and then all the items in the second dictionary, and adding all of them to the new dictionary:
>>> new_context = {} >>> for key, value in context.items(): ... new_context[key] = value ... >>> for key, value in more_context.items(): ... new_context[key] = value ... >>> new_context {'language': 'en', 'timezone': 'UTC', 'title': 'Home', 'breadcrumbs': ['Home']} Merging dictionaries with the update methodWhat if we wanted to âŠ
Read the full article: https://www.pythonmorsels.com/merging-dictionaries/