KEMBAR78
Down the rabbit hole, profiling in Django | PPTX
Down the rabbit hole


     Know your code

     Profiling python
Get those hands dirty
So……
The car engine analogy
Python in that analogy
Scaling
Doing a lot of slow thing at the
          same time
Profiling
An act of balancing
Types of profiling
• (code) profiling (cpu/IO bound problems)
• memory profiling (obviously memory related
  problems)
(Code) profiling
Tools
•   cProfile
•   profile
•   hotshot (deprecated? it splits analysis)
•   line profiler
•   trace
Getting dirty

import cProfile
cProfile.run('foo()')

Or

python -m cProfile myscript.py
More interactive


python -m cProfile myscript.py –o foo.profile

import pstats
p = pstats.Stats('foo.profile')

p.sort_stats('cumulative').print_stats(10)
More complex
import profile             Source: http://www.doughellmann.com/PyMOTW/profile/
def fib(n):
    # from http://en.literateprograms.org/Fibonacci_numbers_(Python)
    if n == 0:
        return 0
    elif n == 1:
        return 1
    else:
        return fib(n-1) + fib(n-2)

def fib_seq(n):
    seq = [ ]
    if n > 0:
        seq.extend(fib_seq(n-1))
    seq.append(fib(n))
    return seq

print 'RAW'
print '=' * 80
profile.run('print fib_seq(20); print')
Output
$ python profile_fibonacci_raw.py
RAW
================================================================================
[0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144, 233, 377, 610, 987, 1597, 2584, 41
81, 6765]

        57356 function calls (66 primitive calls) in 0.746 CPU seconds

  Ordered by: standard name

  ncalls   tottime   percall   cumtime   percallfilename:lineno(function)
      21     0.000     0.000     0.000     0.000:0(append)
      20     0.000     0.000     0.000     0.000:0(extend)
       1     0.001     0.001     0.001     0.001:0(setprofile)
       1     0.000     0.000     0.744     0.744<string>:1(<module>)
       1     0.000     0.000     0.746     0.746profile:0(print fib_seq(20); print)
       0     0.000               0.000          profile:0(profiler)
57291/21     0.743    0.000      0.743    0.035 profile_fibonacci_raw.py:13(fib)
    21/1     0.001    0.000      0.744    0.744 profile_fibonacci_raw.py:22
Output
$ python profile_fibonacci_raw.py
RAW
================================================================================
[0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144, 233, 377, 610, 987, 1597, 2584, 41
81, 6765]

        57356 function calls (66 primitive calls) in 0.746 CPU seconds

  Ordered by: standard name

  ncalls   tottime   percall   cumtime   percallfilename:lineno(function)
      21     0.000     0.000     0.000     0.000:0(append)
      20     0.000     0.000     0.000     0.000:0(extend)
       1     0.001     0.001     0.001     0.001:0(setprofile)
       1     0.000     0.000     0.744     0.744<string>:1(<module>)
       1     0.000     0.000     0.746     0.746profile:0(print fib_seq(20); print)
       0     0.000               0.000          profile:0(profiler)
57291/21     0.743    0.000      0.743    0.035 profile_fibonacci_raw.py:13(fib)
    21/1     0.001    0.000      0.744    0.744 profile_fibonacci_raw.py:22
Memoize!
@memoize
def fib(n):
if n == 0:
        return 0
    elif n == 1:
        return 1
    else:
        return fib(n-1) + fib(n-2)
Yeah!
$ python profile_fibonacci_memoized.py
MEMOIZED
================================================================================
[0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144, 233, 377, 610, 987, 1597, 2584, 41
81, 6765]

        145 function calls (87 primitive calls) in 0.003 CPU seconds

  Ordered by: standard name

  ncalls   tottime   percall   cumtime   percallfilename:lineno(function)
      21     0.000     0.000     0.000     0.000:0(append)
      20     0.000     0.000     0.000     0.000:0(extend)
       1     0.001     0.001     0.001     0.001:0(setprofile)
       1     0.000     0.000     0.002     0.002<string>:1(<module>)
       1     0.000     0.000     0.003     0.003profile:0(print fib_seq(20); print)
       0     0.000               0.000          profile:0(profiler)
   59/21     0.001    0.000      0.001    0.000 profile_fibonacci.py:19(__call__)
      21     0.000    0.000      0.001    0.000 profile_fibonacci.py:26(fib)
    21/1     0.001    0.000      0.002    0.002 profile_fibonacci.py:36(fib_seq)
Line profiling
                                        Source: http://packages.python.org/line_profiler/
Line #      Hits         Time Per Hit    % Time Line Contents
==============================================================
   149                                           @profile
   150                                           def Proc2(IntParIO):
   151     50000        82003      1.6     13.5      IntLoc = IntParIO + 10
   152     50000        63162      1.3     10.4      while 1:
   153     50000        69065      1.4     11.4           if Char1Glob == 'A':
   154     50000        66354      1.3     10.9               IntLoc = IntLoc - 1
   155     50000        67263      1.3     11.1               IntParIO = IntLoc -
IntGlob
   156     50000        65494      1.3     10.8               EnumLoc = Ident1
   157     50000        68001      1.4     11.2           if EnumLoc == Ident1:
   158     50000        63739      1.3     10.5               break
   159     50000        61575      1.2     10.1      return IntParIO
More complex == visualize


Run
Snake
Run
Kcachegrind


Pyprof
    2
calltree
What to look for?
• Things you didn’t expect ;)
• much time spend in one function
• lots of calls to same function
Performance solutions
• caching
• getting stuff out of inner loops
• removing logging
Memory profiling


Django has a memory leak

    In debug mode ;)
Tools
• heapy (and pysizer)
• meliea (more like hotshot)
Meliea!
• Memory issues happen sometimes in
  production (long living processes). Want to get
  info, then analyze locally
• runsnakerun has support for it
Dumping
 Source: http://jam-bazaar.blogspot.com/2009/11/memory-debugging-with-meliae.html




from meliae import scanner
scanner.dump_all_objects('filename.json')
Analyzing
>>> from meliae import loader
>>> om = loader.load('filename.json')
>>> s = om.summarize(); s

This dumps out something like:
Total 17916 objects, 96 types, Total size = 1.5MiB (1539583 bytes)
Index   Count    %     Size   % Cum     Max Kind
   0     701   3    546460 35 35     49292 dict
   1    7138 39     414639 26 62      4858 str
   2     208   1     94016   6 68      452 type
   3    1371   7     93228   6 74       68 code
   4    1431   7     85860   5 80       60 function
   5    1448   8     59808   3 84      280 tuple
   6     552   3     40760   2 86      684 list
   7      56   0     29152   1 88      596 StgDict
   8    2167 12      26004   1 90       12 int
   9     619   3     24760   1 91       40 wrapper_descriptor
  10     570   3     20520   1 93       36 builtin_function_or_method
  ...
Run s n a k   e   run
Your production system


  A completely different story
What can you do to prevent this?
• From an API perspective:
  – Maybe one WSGI process that is different
  – Gets a small amount of requests (load balancing)
  – This process takes care of doing profiling
    (preferably in the hotshot way)
Or…..
• Pycounters

• Author is in tha house: Boaz Leskes
Finally
• You should know about this
• Part of your professional toolkit

• This should be in IDE’s!
  – Komodo already has it, what about
    PyCharm??(can you blog this Reinout? ;)
Questions?
Links
Articles:
http://www.doughellmann.com/PyMOTW/profile/
http://www.doughellmann.com/PyMOTW/trace/
http://jam-bazaar.blogspot.com/2010/08/step-by-step-meliae.html
https://code.djangoproject.com/wiki/ProfilingDjango

Videos:
http://www.youtube.com/watch?v=Iw9-GckD-gQ
http://blip.tv/pycon-us-videos-2009-2010-2011/introduction-to-python-profiling-1966784

Software:
http://www.vrplumber.com/programming/runsnakerun/
http://kcachegrind.sourceforge.net/html/Home.html
http://pypi.python.org/pypi/pyprof2calltree/
https://launchpad.net/meliae
http://pypi.python.org/pypi/line_profiler
http://pypi.python.org/pypi/Dozer

Down the rabbit hole, profiling in Django

Editor's Notes

  • #4 Introduction: - how many of you do profiling - those who don&apos;t, who would say they do more complex projects? - what do you do? - how well do you know your code?I&apos;m sorry but you don&apos;t
  • #6 Python has helped us tremendously to think on an abstract level about the problems we solve. I totally believe in being less worried about how code gets executed on a low level. Because most of the time I just don&apos;t care (TM). Why don&apos;t I care? Even though I&apos;m an engineer, I care about solving problems for people (and making a buck out of it). We do Recharted, which means that our clients want to give their customers the best experience possible. That means we should return requested info in a timely manner. Ever had that you where waiting irritated because of waiting for gmail?We had an issue (talked about during my logging presentation) with SUDS soap library doing a LOT of logging debug calls. Which on a call to a SOAP api for booking flights costed us 10-15 seconds on complex responses!
  • #7 Scalability vs. performanceWe hear a lot about scaling, but sometimes we forget performance. Scalability means you can do the same thing for a lot of people. And that more people has a small impact on your performance. But that still means you can have the same shitty baseline perfomance. Actually it is not at all hard to scale shitty perfomance almost :)
  • #8 You can make time.sleep() scale very well (with the right server infrastructure of course)
  • #9 So what is profiling.Basically profiling is running your code in the interpreter, but in a way that statistics are recorded during the actual run. (yes this has a performance impact, so you can&apos;t just do this in production. For that there are other ways). Then you look at those statistics. This actually gives you a lot of insight in what happens.I know what happens in the local scope? But what actually happens when an API wsgi request comes in until we deliver the response? You would actually be surprised to know how much stuff happens in between. This is part of actually getting to know your code. Btw your code isn&apos;t just you, what about libraries you use? Systems you interface with? This all has an impact on your performance.
  • #10 Actually profiling allows you to zoom in on low hanging fruit, you should ALWAYS balance amount of work/changing code vs. relative wins in performance. Basically you&apos;ll most often find fixing the two top entries :)
  • #17 Ncallsfor the number of calls,Tottimefor the total time spent in the given function (and excluding time made in calls to sub-functions),Percallis the quotient of tottime divided by ncallsCumtimeis the total time spent in this and all subfunctions (from invocation till exit). This figure is accurate even for recursive functions.Percallis the quotient of cumtime divided by primitive callsfilename:lineno(function)provides the respective data of each function
  • #18 Ncallsfor the number of calls,Tottimefor the total time spent in the given function (and excluding time made in calls to sub-functions),Percallis the quotient of tottime divided by ncallsCumtimeis the total time spent in this and all subfunctions (from invocation till exit). This figure is accurate even for recursive functions.Percallis the quotient of cumtime divided by primitive callsfilename:lineno(function)provides the respective data of each function
  • #20 Ncallsfor the number of calls,Tottimefor the total time spent in the given function (and excluding time made in calls to sub-functions),Percallis the quotient of tottime divided by ncallsCumtimeis the total time spent in this and all subfunctions (from invocation till exit). This figure is accurate even for recursive functions.Percallis the quotient of cumtime divided by primitive callsfilename:lineno(function)provides the respective data of each function
  • #26 Just because a language is garbage collected doesn&apos;t mean it can&apos;t leak memory - C modules could leak (harder to find) - Some globally available datastructure could live and grow without you knowing (actually django has a builtin memory leak while running in debug mode) - Cyclic reference funky stuff, letting the interpreter think that memory cannot be released.Also (like with profiling), knowing the memory profile of your application could help. Maybe you have application server instances that are 64 mb. If a quarter of that is unneccasary stuff, you could maybe run more instances on the same hardware! Leading towards faster world domination!
  • #32 So now you find out that your code behaves beautifully on your local machine. And then when in production... borkbork.
  • #35 In conclusion:I think every serious python developer should know about these things, it is part of your toolkit.