- The author created a system that can generate an entire font set from a single sample image by extracting the font style and generating images based on the extracted style and class information.
- The system uses a style extractor CNN to extract the style vector from an input image and a GlyphGAN generator to generate images conditioned on the extracted style vector and class vector.
- Evaluation shows the generated images have reasonable legibility but the extracted style vector does not fully capture consistency, with average similarity to the training dataset around 50-60%. Future work aims to improve the style extraction accuracy.
AN INTRODUCTION TO AUTO-ML EDGE-ML (VIDEO 1/4)Alexis Bondu
The classical approches of Auto-ML automatise the best practices of Data Scientists : they assess a great number of models and select the most precise.
Edge-ML offers a real break through thanks to MODL approch that allows to produce efficient models in a simple way with warranted robustness.
A seminar in advanced Software Engineering concerning using models to guide the development process, and QVT to transfer a model into another model automatically
Machine learning lets you make better business decisions by uncovering patterns in your consumer behavior data that is hard for the human eye to spot. You can also use it to automate routine, expensive human tasks that were previously not doable by computers. In the business to business space (B2B), if your competitors can make wiser business decisions based on data and automate more business operations but you still base your decisions on guesswork and lack automation, you will lose out on business productivity. In this introduction to machine learning tech talk, you will learn how to use machine learning even if you do not have deep technical expertise on this technology.
Topics covered:
1.What is machine learning
2.What is a typical ML application architecture
3.How to start ML development with free resource links
4.Key decision factors in ML technology selection depending on use case scenarios
AN INTRODUCTION TO AUTO-ML EDGE-ML (VIDEO 1/4)Alexis Bondu
The classical approches of Auto-ML automatise the best practices of Data Scientists : they assess a great number of models and select the most precise.
Edge-ML offers a real break through thanks to MODL approch that allows to produce efficient models in a simple way with warranted robustness.
A seminar in advanced Software Engineering concerning using models to guide the development process, and QVT to transfer a model into another model automatically
Machine learning lets you make better business decisions by uncovering patterns in your consumer behavior data that is hard for the human eye to spot. You can also use it to automate routine, expensive human tasks that were previously not doable by computers. In the business to business space (B2B), if your competitors can make wiser business decisions based on data and automate more business operations but you still base your decisions on guesswork and lack automation, you will lose out on business productivity. In this introduction to machine learning tech talk, you will learn how to use machine learning even if you do not have deep technical expertise on this technology.
Topics covered:
1.What is machine learning
2.What is a typical ML application architecture
3.How to start ML development with free resource links
4.Key decision factors in ML technology selection depending on use case scenarios
Lessons Learned from Building Machine Learning Software at NetflixJustin Basilico
Talk from Software Engineering for Machine Learning Workshop (SW4ML) at the Neural Information Processing Systems (NIPS) 2014 conference in Montreal, Canada on 2014-12-13.
Abstract:
Building a real system that incorporates machine learning as a part can be a difficult effort, both in terms of the algorithmic and engineering challenges involved. In this talk I will focus on the engineering side and discuss some of the practical issues we’ve encountered in developing real machine learning systems at Netflix and some of the lessons we’ve learned over time. I will describe our approach for building machine learning systems and how it comes from a desire to balance many different, and sometimes conflicting, requirements such as handling large volumes of data, choosing and adapting good algorithms, keeping recommendations fresh and accurate, remaining responsive to user actions, and also being flexible to accommodate research and experimentation. I will focus on what it takes to put machine learning into a real system that works in a feedback loop with our users and how that imposes different requirements and a different focus than doing machine learning only within a lab environment. I will address the particular software engineering challenges that we’ve faced in running our algorithms at scale in the cloud. I will also mention some simple design patterns that we’ve fond to be useful across a wide variety of machine-learned systems.
How to use transfer learning to bootstrap image classification and question a...Wee Hyong Tok
#theaiconf SFO 2018
Session by Danielle Dean, WeeHyong Tok
Transfer learning enables you to use pretrained deep neural networks trained on various large datasets (ImageNet, CIFAR, WikiQA, SQUAD, and more) and adapt them for various deep learning tasks (e.g., image classification, question answering, and more).
Wee Hyong Tok and Danielle Dean share the basics of transfer learning and demonstrate how to use the technique to bootstrap the building of custom image classifiers and custom question-answering (QA) models. You’ll learn how to use the pretrained CNNs available in various model libraries to custom build a convolution neural network for your use case. In addition, you’ll discover how to use transfer learning for question-answering tasks, with models trained on large QA datasets (WikiQA, SQUAD, and more), and adapt them for new question-answering tasks.
https://conferences.oreilly.com/artificial-intelligence/ai-ca/public/schedule/detail/68527
Presentation on BornoNet Research Paper and Python BasicsShibbir Ahmed
The slides are of a presentation on BornoNet Research Paper and Python basics done by our team recently in our Mobile and Telecommunication course of undergraduate studies.
Automated machine learning lectures given at the Advanced Course on Data Science & Machine Learning. AutoML, hyperparameter optimization, Bayesian optimization, Neural Architecture Search, Meta-learning, MAML
Lessons Learned from Building Machine Learning Software at NetflixJustin Basilico
Talk from Software Engineering for Machine Learning Workshop (SW4ML) at the Neural Information Processing Systems (NIPS) 2014 conference in Montreal, Canada on 2014-12-13.
Abstract:
Building a real system that incorporates machine learning as a part can be a difficult effort, both in terms of the algorithmic and engineering challenges involved. In this talk I will focus on the engineering side and discuss some of the practical issues we’ve encountered in developing real machine learning systems at Netflix and some of the lessons we’ve learned over time. I will describe our approach for building machine learning systems and how it comes from a desire to balance many different, and sometimes conflicting, requirements such as handling large volumes of data, choosing and adapting good algorithms, keeping recommendations fresh and accurate, remaining responsive to user actions, and also being flexible to accommodate research and experimentation. I will focus on what it takes to put machine learning into a real system that works in a feedback loop with our users and how that imposes different requirements and a different focus than doing machine learning only within a lab environment. I will address the particular software engineering challenges that we’ve faced in running our algorithms at scale in the cloud. I will also mention some simple design patterns that we’ve fond to be useful across a wide variety of machine-learned systems.
How to use transfer learning to bootstrap image classification and question a...Wee Hyong Tok
#theaiconf SFO 2018
Session by Danielle Dean, WeeHyong Tok
Transfer learning enables you to use pretrained deep neural networks trained on various large datasets (ImageNet, CIFAR, WikiQA, SQUAD, and more) and adapt them for various deep learning tasks (e.g., image classification, question answering, and more).
Wee Hyong Tok and Danielle Dean share the basics of transfer learning and demonstrate how to use the technique to bootstrap the building of custom image classifiers and custom question-answering (QA) models. You’ll learn how to use the pretrained CNNs available in various model libraries to custom build a convolution neural network for your use case. In addition, you’ll discover how to use transfer learning for question-answering tasks, with models trained on large QA datasets (WikiQA, SQUAD, and more), and adapt them for new question-answering tasks.
https://conferences.oreilly.com/artificial-intelligence/ai-ca/public/schedule/detail/68527
Presentation on BornoNet Research Paper and Python BasicsShibbir Ahmed
The slides are of a presentation on BornoNet Research Paper and Python basics done by our team recently in our Mobile and Telecommunication course of undergraduate studies.
Automated machine learning lectures given at the Advanced Course on Data Science & Machine Learning. AutoML, hyperparameter optimization, Bayesian optimization, Neural Architecture Search, Meta-learning, MAML
Similar to May internship challenge: Font Generator (20)
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
2. Overview
I created a system that can extract style information from a single sample
image and generate an entire font set with uniformity.
2
Output: Image of the entire font set
Input: 1 font image
3. Table of contents
• Background
• Method
• Results and discussion
• Future work
3
4. Background: Creating a font set
Creating a font set is very labor intensive.
The only way is for the creator to prepare the images one by one.
4
The more characters there are,
the longer it takes to create.
If it takes 3 min. /character…
Alphabet: {A, B, C, …}
→ 52 classes = 2.6 h
Kanji: {一, 二, 三, …}
→ 2136 classes = 106.8 h
5. Background: Style consistency of font sets
It is not necessary to check all the fonts to get a feel for the font set.
5
Font samples. You can get an idea of the overall atmosphere from just some of the letters.
Avenir Next
Lorem ipsum dolor sit amet,
consectetur adipiscing elit,
sed do eiusmod tempor
incididunt ut labore et
dolore magna aliqua.
Baskerville
Lorem ipsum dolor sit
amet, consectetur
adipiscing elit, sed do
eiusmod tempor incididunt
ut labore et dolore magna
aliqua.
Didot
Lorem ipsum dolor sit amet,
consectetur adipiscing elit,
sed do eiusmod tempor
incididunt ut labore et
dolore magna aliqua.
6. Problem-setting: Generating an entire font set from a few samples
Considering the problem of generating an entire unified font set from a subset
of samples in the font set.
6
Font sample
ABCDEFG
HIJKLMN
OPQRSTU
VWXYZ
Entire font set
7. Hypothesis: How to generate an entire font set from a few samples?
It would be effective to extract font styles from a few samples and
generate them based on the font style and class information.
7
Font sample
ABCDEFG
HIJKLMN
OPQRSTU
VWXYZ
A
Entire font set
𝑧𝑠
Style information.
Mincho or Gothic, etc.
𝑧𝑐
Class information. A, B, … etc.
Extraction
8. Table of contents
• Background
• Method
• Results and discussion
• Future work
8
10. Generator: GlyphGAN (1/4)
GlyphGAN (Hayashi et al., 2019) is a type of GAN that generates
a consistent and diverse font sets.
10
𝑧
𝑧𝑠
𝑧𝑐
𝐺(𝑧)
𝑥
𝐷
Dataset
Generated Images
Generator
Input
Style Vector
Class Vector
Discriminator
𝑦
𝐺
11. Generator: GlyphGAN (2/4)
The generator and discriminator are CNN-based models.
11
Generator (top) and Discriminator (bottom)
Photo by (Hayashi et al., 2019)
12. Generator: GlyphGAN (3/4)
Input consists of style information and class information.
12
𝑧
𝑧𝑠
𝑧𝑐
Input
Style Vector
Class Vector
𝑧𝑐
: Class Vector
ex.
A → [1,0, ⋯ , 0]𝑇
B → [0,1, ⋯ , 0]𝑇
⋮
Z → [0,0, ⋯ , 1]𝑇
𝑧𝑠: Style Vector
𝑧𝑠
∈ ℝ𝑛
, 𝑧𝑖
𝑠
∼ 𝑈(−1,1)
ex.
[0.1, −0.7, ⋯ , 0.5] ∈ ℝ100
14. Extractor: CNN-based model (1/2)
Outputs a style vector with a single sample image as input.
14
Style Extractor
𝐸 𝑧
Output
𝑧𝑠
Style Vector
𝑥
Image
The structure of the style extractor
The structure is the same as
GlyphGAN’s Discriminator
except for the last layer.
Photo by (Hayashi et al., 2019)
15. Extractor: CNN-based model (2/2)
Create a dataset using a trained generator.
15
𝑧 𝐺(𝑧)
Dataset for training
Style Extractor
𝐸 𝑧
Output
𝑧𝑠
Style Vector
𝑥
Image
16. Table of contents
• Background
• Method
• Results and discussion
• Future work
16
17. Generator: Training dataset
Dataset: Alphabet Characters Fonts Dataset
Number of data: 26 classes × 6561 font types
17
Sample images in the dataset
18. Generator: Training
Training settings
• batch size: 1024
• epochs: 2500
• optimizer: Adam (lr=0.0002)
• criterion: WGAN-GP
18
Learning curve (Wasserstein distance)
22. Examples of image generation
22
Output: Image of the entire font set
Input: 1 font image
Style extraction
& Generation
Style extraction
& Generation
23. Evaluation 1: Legibility of generated images (1/3)
Create a CNN-based multi-class classification model.
Compare the accuracy on the dataset with that on the generated images.
23
The structure of multi-font classifier
Photo by (Hayashi et al., 2019)
24. Evaluation 1: Legibility of generated images (2/3)
Number of data:
train: 26 classes × 6561 font types,
validation: 26 classes × 8429 font types
Training settings:
• batch size: 1024
• epochs: 100
• optimizer: Adam (lr=0.0002)
• criterion: cross entropy
24
Learning curves: accuracy (top) and loss (bottom)
25. Evaluation 1: Legibility of generated images (3/3)
Evaluation results
It can be confirmed that a certain level of readability has been achieved.
25
Accuracy
Training dataset (6561 font types) 97.0%
Test dataset (8429 font types) 89.9%
Generated fonts
(10000 font types)
82.6%
26. Evaluation 2: Style extraction (1/3)
Calculate the average similarity (SSIM) between the fonts in the dataset and the
fonts generated by style extraction and generation.
26
Fonts in dataset
Style extraction
& Generation
Generated fonts
Calculation of similarity (SSIM)
27. Evaluation 2: Style extraction (2/3)
SSIM is a perception-based model that considers image degradation as
perceived change in structural information.
27
MSE vs. SSIM
Photo by (Wang and Bovik, 2009)
𝑆𝑆𝐼𝑀(𝑥, 𝑦) =
(2𝜇𝑥𝜇𝑦 + 𝑐1)(2𝜎𝑥𝑦 + 𝑐2)
(𝜇𝑥
2
+ 𝜇𝑦
2
+ 𝑐1)(𝜎𝑥
2
+ 𝜎𝑦
2
+ 𝑐2)
28. Evaluation 2: Style extraction (3/3)
One character from each font set was randomly selected.
Evaluation results
Style extraction works to some extent, but not well enough,
28
Average similarity
Training dataset (6561 font types) 67.2%
Test dataset (8429 font types) 66.0%
29. Evaluation 3: Style consistency (1/2)
Calculate the average similarity (SSIM) between the font sets in the dataset and
the font sets generated by style extraction and generation.
29
Font set in dataset
ABCDEFG
HIJKLMN
OPQRSTU
VWXYZ
A
Sampling
Style extraction
& Generation
ABCDEFG
HIJKLMN
OPQRSTU
VWXYZ
Generated font set
Calculation of similarity (SSIM)
30. Evaluation 3: Style consistency (2/2)
Evaluation results
The similarity between font sets is not high.
The low accuracy of the extractor may be a bottleneck.
30
Average similarity
Training dataset (6561 font types) 52.4%
Test dataset (8429 font types) 51.4%
32. Table of contents
• Background
• Method
• Results and discussion
• Future work
32
33. Current problem 1: Accuracy of the extractor
Improvements could be made by using the SSIM losses of the original and
generated images during training.
33
Style Extractor
𝐸
𝑥
Image
𝑧 𝐺(𝑧)
Generated Images
Generator
Input
𝐺
34. Current problem 2: Inefficiency of extractor training
After the generator is trained, the corresponding extractor needs to be trained.
This could be improved by using models such as VAE or flow-based models.
34
VAE (top) and Flow-based model (bottom)
Encoder
𝐸
𝑥
Image
𝑧 𝐷(𝑧)
Generated Images
Decoder
Latent
𝐷
Flow
𝑓
𝑥
Image
𝑧 𝑓−1(𝑧)
Generated Images
Inverse
𝑓−1
Latent
35. Current problem 3: Small dataset
The relationship between the number of datasets and the accuracy of
generation needs to be investigated.
35
Results for Hiragana dataset
Number of data: 84 classes × 50 font types
36. Application examples of Style Extractor + GlyphGAN System
If enough datasets can be prepared, applications that reduce the burden on
creators can be considered.
36
ex. A system to create assets of your own art style from a single sample.
37. Conclusion
What I made:
A system that combines Style Extractor and GlyphGAN to create an entire font
set from a single font image.
Level of achievement:
• A certain level of readability.
• The reproduction of style remains an issue.
37
38. References
• [1] Hayashi et al., “GlyphGAN: Style-Consistent Font Generation Based on
Generative Adversarial Networks”, 2019.
• [2] Wang and Bovik, “Mean squared error: Love it or leave it? A new look at
Signal Fidelity Measures”, 2009.
38