Upgrade to Pro
— share decks privately, control downloads, hide ads and more …
Speaker Deck
Features
Speaker Deck
PRO
Sign in
Sign up for free
Search
Search
An Asynchronous, Scalable Django with Twisted (...
Search
Amber Brown (HawkOwl)
June 04, 2016
Programming
0
190
An Asynchronous, Scalable Django with Twisted (PyCon TW 2016 Keynote)
Amber Brown (HawkOwl)
June 04, 2016
Tweet
Share
More Decks by Amber Brown (HawkOwl)
See All by Amber Brown (HawkOwl)
Batteries Included, But They're Leaking (Python Language Summit, 2019)
hawkowl
0
480
Why Twisted Is The Best (WOOTConf @ LCA 2017)
hawkowl
0
150
Releasing Calendar Versioned Software (PyCon AU 2016)
hawkowl
1
140
Falsehoods Developers Have About Identity (PyCon AU 2016 Lightning Talk)
hawkowl
1
780
The Report of Twisted's Death; or Twisted & Tornado in the Asyncio Age (EuroPython 2016)
hawkowl
0
150
The Report of Twisted's Death; or Twisted & Tornado in the Asyncio Age (PyCon US 2016)
hawkowl
0
180
Twisted & Python 3 (Python Language Summit, 2016)
hawkowl
0
170
The Report of Twisted's Death; or Twisted & Tornado in the Asyncio Age (Perth Django Meetup)
hawkowl
0
160
The Future of Twisted, and Pretty Much Everything Else (PyCon CZ Keynote, 2015)
hawkowl
0
130
Other Decks in Programming
See All in Programming
11年かかって やっとVibe Codingに 時代が追いつきましたね
yimajo
1
240
なぜ今、Terraformの本を書いたのか? - 著者陣に聞く!『Terraformではじめる実践IaC』登壇資料
fufuhu
4
410
中級グラフィックス入門~効率的なメッシュレット描画~
projectasura
4
2.5k
Flutterと Vibe Coding で個人開発!
hyshu
1
230
一人でAIプロダクトを作るための工夫 〜技術選定・開発プロセス編〜 / I want AI to work harder
rkaga
4
470
実践!App Intents対応
yuukiw00w
1
200
知って得する@cloudflare_vite-pluginのあれこれ
chimame
1
140
React 使いじゃなくても知っておきたい教養としての React
oukayuka
18
5.3k
DMMを支える決済基盤の技術的負債にどう立ち向かうか / Addressing Technical Debt in Payment Infrastructure
yoshiyoshifujii
5
760
なぜあなたのオブザーバビリティ導入は頓挫するのか
ryota_hnk
5
570
DataformでPythonする / dataform-de-python
snhryt
0
150
「リーダーは意思決定する人」って本当?~ 学びを現場で活かす、リーダー4ヶ月目の試行錯誤 ~
marina1017
0
130
Featured
See All Featured
Intergalactic Javascript Robots from Outer Space
tanoku
272
27k
Learning to Love Humans: Emotional Interface Design
aarron
273
40k
個人開発の失敗を避けるイケてる考え方 / tips for indie hackers
panda_program
110
19k
Scaling GitHub
holman
461
140k
Fight the Zombie Pattern Library - RWD Summit 2016
marcelosomers
234
17k
Distributed Sagas: A Protocol for Coordinating Microservices
caitiem20
332
22k
Templates, Plugins, & Blocks: Oh My! Creating the theme that thinks of everything
marktimemedia
31
2.5k
The Cost Of JavaScript in 2023
addyosmani
51
8.8k
What’s in a name? Adding method to the madness
productmarketing
PRO
23
3.6k
Why You Should Never Use an ORM
jnunemaker
PRO
58
9.5k
Let's Do A Bunch of Simple Stuff to Make Websites Faster
chriscoyier
507
140k
Balancing Empowerment & Direction
lara
1
530
Transcript
An Asynchronous, Scalable Django with Twisted
Hello, I’m Amber Brown (HawkOwl)
Twitter: @hawkieowl Pronouns: she/her
I live in Perth, Western Australia
None
Core Developer Release Manager Ported 40KLoC+ to Python 3
None
Binary release management across 3 distros Ported Autobahn|Python (Tx) and
Crossbar.io to Python 3 Web API/REST integration in CB
Scaling Django Applications
Django serves one request at a time
gunicorn, mod_wsgi, etc run multiple copies in threads + processes
Concurrent Requests == processes x threadpool size
nginx gunicorn worker thread thread thread thread gunicorn worker thread
thread thread thread Example server: two workers with four threads each
Need more requests? Add more web servers!
nginx gunicorn worker thread thread thread thread gunicorn worker thread
thread thread thread nginx gunicorn worker thread thread thread thread gunicorn worker thread thread thread thread HAProxy Server 2 Server 3 Server 1
Scaling has required adding a new piece
Higher scale means higher complexity
Is there a better way to handle many requests?
Problem Domain
Modern web applications have two things that take a long
time to do
CPU-bound work Math, natural language processing, other data processing
On most Python interpreters, Python threads are unsuitable for dispatching
CPU-heavy work
Of N Python threads only 1 may run Python code
because of the Global Interpreter Lock
Of N Python threads only N may run C code,
since the Global Interpreter Lock is released
I/O-bound work Database requests, web requests, other network I/O
Threads work better for I/O-bound work
Thread switching overhead is expensive Rapidly acquiring/releasing the GIL is
expensive
First, let's focus on I/O-bound applications.
Asynchronous I/O & Event-Driven Programming
Your code is triggered on events
Events can be: incoming data on the network some computation
is finished a subprocess has ended etc, etc
How do we know when events have occurred?
All events begin from some form of I/O, so we
just wait for that!
Event-driven programming frameworks
Twisted (the project I work on!)
(of SVN history)
asyncio was introduced much later
None
Same at their core, using "selector functions"
select() and friends (poll, epoll, kqueue)
Selector functions take a list of file descriptors (e.g. sockets,
open files) and tell you what is ready for reading or writing
Selector loops can handle thousands of open sockets and events
Data is channeled through a transport to a protocol (e.g.
HTTP, IMAP, SSH)
Sending data is queued until the network is ready
Nothing blocks, it simply gives control to the next event
to be processed
No blocking means no threads
“I/O loops” or “reactors” (as it "reacts" to I/O)
Higher density per core No threads required! Concurrency, not parallelism
Best case: high I/O throughput, high-latency clients, low CPU processing
But what if we need to process CPU bound tasks?
Event Driven Programming with Work Queues
CPU bound tasks are added to a queue, rather than
being ran directly
Web Server Task Queue Worker Worker Worker
We have made the CPU-bound task an I/O-bound one for
our web server
We have also made the scaling characteristics horizontal
Web Server Task Queue Worker Server 2 CPU3 Worker Server
2 CPU2 Worker Server2 CPU1 Worker Server 1 CPU2 Worker Server 1 CPU1 Worker Server 2 CPU4
Putting tasks on the queue and removing them is cheap
Task queues scale rather well
Add more workers to scale!
Do we have an implementation of this?
The Architecture of Django Channels
Project to make an "asynchronous Django"
Authored by Andrew Godwin (behind South, Migrations)
Interface Server Channel Queue Worker Worker Worker Worker Worker Worker
Server 1 Server 2 Server 3 Server 4
Interface server accepts requests, puts them on the Channel (task
queue)
Workers take requests off the Channel and process them
Results from processed requests are written back to the Channel
The interface server picks up these responses and writes it
back out to the HTTP request
The interface server is only I/O bound and does no
"work" of its own
Perfect application for asynchronous I/O!
Daphne, the reference interface server implementation, is written in Twisted
Daphne is capable of handling thousands of requests a second
on modest hardware
The channel layer can be sharded
Channel Queue Server 2 Interface Server Worker Worker Worker Worker
Worker Worker Server 1 Server 4 Server 5 Channel Queue Server 3 (Sharding)
Workers do not need to be on the web server...
but you can put them there if you want!
For small sites, the channel layer can simply be an
Inter- Process-Communication bus
Channel Queue (Shared Memory) Interface Server Worker Worker Worker Server
1
And Twisted understands WebSockets... so can Channels too?
Yep!
How Channels Works
A Channel is where requests are put to be serviced
What is a request? - incoming HTTP requests - connected
WebSocket connection - data on a WebSocket
http.request http.disconnect websocket.connect websocket.receive websocket.disconnect
Your worker listens on these channel names
Information about the request (e.g. a body and headers), and
a "reply channel" identifier
http.response!<client> http.request.body!<client> websocket.send!<client>
http.response!c134x7y http.request.body!c134x7y websocket.send!c134x7y
Reply channels are connection specific so that the correct response
gets to the correct connection
In handling a request, your code calls send() on a
response channel
But because Channels is event-driven, you can't get a "response"
from the event
The workers themselves do not use asynchronous I/O by default!
Under Channels, you write synchronous code, but smaller synchronous code
@receiver(post_save, sender=BlogUpdate) def send_update(sender, instance, **kwargs): Group("liveblog").send({ "id": instance.id, "content":
instance.content, })
Group? What's a group?
Pool of request-specific channels for efficiently sending one-to-many messages
e.g: add all open WebSocket connections to a group that
is notified when your model is saved
Handling different kinds of requests
Workers can listen on specific channels, they don't have to
listen to all of them!
Interface Server Channel Queue Worker Worker Server 1 Server 2
Server 3 (high performance) Server 4 (standard) http.request bigdata.process
Because you can create and listen for arbitrary channels, you
can funnel certain kinds of work into different workers
my_data_set = request.body Channel("bigdata.process").send( {"mydata": my_data_set})
How do we support sending that data down the current
request when it's done?
my_data_set = request.body Channel("bigdata.process").send({ "mydata": my_data_set, "reply_channel": message.reply_channel})
All our big data worker needs to do then is
send the response on the reply channel!
Channels as a bridge to an asynchronous future
A channel doesn't care if you are synchronous or asynchronous
...or written in Django or even Python!
Channels implements "Asynchronous Server Gateway Interface"
The path to a hybrid future Go, Django, Twisted, etc
etc
Channels is due to land in Django 1.11/2.0
Try it out! channels.readthedocs.io
Questions? (pls no statements, save them for after)