Differences between revisions 4 and 10 (spanning 6 versions)
Revision 4 as of 2006-08-10 09:41:17
Size: 3396
Comment:
Revision 10 as of 2007-01-25 15:47:33
Size: 3504
Comment:
Deletions are marked like this. Additions are marked like this.
Line 5: Line 5:
== Part of my Thesis ==
 * My thesis deals with algebraic attacks on block ciphers namely the Courtois Toy Cipher. So I will implement/have implemented several algebraic attack algorithms like XL,XSL,F4 and DR. Though those might not be of general interest.
== Other stuff I'm working on ==
 * Cputime class/function which wraps all the cputime() calls for all the subprocesses for convenience. So only one '''cputime(all=True)''' call would be sufficient.


== Stuff I'm working on ==

 * ntl.GF2E & ntl.ZZpX wrapping for finite fields

 * Speed up MPolynomials by linking to Singular (this is going to be an exciting experience on its own) or CoCoALib

 * Possible: Specialized MPolynomialGF2 class

 * Possible: Improve F4 implementation or start over in cooperation with some of the other developers working on F4
right now

 * Clean-up Ideal class (make sure a basis != ideal)
Line 11: Line 21:
Line 12: Line 23:
 * Allow %prun and hotshot to profile pyrex code.
 * ["Givaro"] integration. This significantly improves finite field arithmetic (more than ten times) and everything relying on it in SAGE. Try it: '''k = linbox.GFq(2^8,repr="poly")'''
 * Extremely fast sparse (and dense) linear algebra. Actually I only care about echelon form calculation. We are debating if [http://www.linalg.org LinBox] is a good choice for this. Matrices involved in algebraic attacks are often very sparse and there is no need to have my machine go down because e.g., NTL allocates many zeros (i.e., NTL only knows dense matrices).
Line 18: Line 27:
Line 19: Line 29:
Line 20: Line 31:
 * NTL wrapper makeover (more SAGEish, avoid function calls, avoid news, deletes), CF wrapper makeover (avoid function calls, avoid news,deletes)
* NTL wrapper makeover (more SAGEish, avoid function calls, avoid news, deletes)
Line 22: Line 35:

== Done ==

== Part of my Thesis (Diplom) ==
 * My thesis deals with algebraic attacks on block ciphers namely the Courtois Toy Cipher. So I implemented several algebraic attack algorithms like XL, F4, and DR and a slightly optimized MPolynomial over GF(2) class. Though those might not be of general interest. I will push some of that stuff upstream.
 * Givaro is going to be in SAGE 1.5
 * Cputime class/function which wraps all the cputime() calls for all the subprocesses for convenience. So only one '''cputime(all=True)''' call would be sufficient. (I extended David Harvey's Profiler class for this)
Line 35: Line 55:
}}} This is due to the way multivariate polynomials in SAGE are represented. I want to come up with a more sparse representation which does not add zero to zero 1998 times in the second example. }}} This is due to the way multivariate polynomials in SAGE are represented. I want to come up with a more sparse representation which does not add zero to zero 1998 times in the second example. (I rewrote the polynomial representation to use dicts of dicts which map indices to exponents e.g., {{1:2}:3} represents 3*y^2 if y is the second variable in the ring.)
Line 37: Line 57:
= Other stuff =
=
= Other stuff ==
Line 39: Line 60:

Martin Albrecht's (malb) SAGE projects

Stuff I'm working on

  • ntl.GF2E & ntl.ZZpX wrapping for finite fields

  • Speed up MPolynomials by linking to Singular (this is going to be an exciting experience on its own) or CoCoALib
  • Possible: Specialized MPolynomialGF2 class
  • Possible: Improve F4 implementation or start over in cooperation with some of the other developers working on F4

right now

  • Clean-up Ideal class (make sure a basis != ideal)
  • Memory consumption analogous to cputime(). This is tricky because some grep on top et al. doesn't provide the necessary information. E.g. Python never ever frees memory while running, and we also might count shared libraries several times this way.
  • Memory profiling similar to %prun or hotshot. The memory profiler would provide hints which part consumes the most memory during a calculation.
  • Parallel sparse linear algebra to utilize all 16 cores at sage.math.washington.edu at once. :) It's either too hard for me (if there is no library) or simple as it would be just another library to expose.

  • [http://article.gmane.org/gmane.comp.mathematics.sage.general/193/ SAGEBot] is not dead yet.

  • [http://eprint.iacr.org/2006/224.pdf Generalizations of the Karatsuba Algorithm for Efficient Implementations]

  • Implement or wrap the [http://eprint.iacr.org/2006/251.pdf "Method of Four Russians"] for row reducing resp. inverting a dense boolean matrix

  • NTL wrapper makeover (more SAGEish, avoid function calls, avoid news, deletes)
  • [:Factory:cf.CF] makeover (avoid function calls, avoid news, deletes). Also restrict arithmetic to cf.CF (i.e. strongly type it) to avoid casting/coercion overhead.

Done

Part of my Thesis (Diplom)

  • My thesis deals with algebraic attacks on block ciphers namely the Courtois Toy Cipher. So I implemented several algebraic attack algorithms like XL, F4, and DR and a slightly optimized MPolynomial over GF(2) class. Though those might not be of general interest. I will push some of that stuff upstream.
  • Givaro is going to be in SAGE 1.5
  • Cputime class/function which wraps all the cputime() calls for all the subprocesses for convenience. So only one cputime(all=True) call would be sufficient. (I extended David Harvey's Profiler class for this)

  • Consider this example:

       1 sage: R1 = PolynomialRing(GF(2**8),2)
       2 sage: R2 = PolynomialRing(GF(2**8),2000)
       3 sage: x1=R1.gen()
       4 sage: y1=R1.gen(1)
       5 sage: x2=R2.gen()
       6 sage: y2=R2.gen(1)
       7 sage: time for i in range(1000): _ =x2*y2
       8 CPU times: user 1.58 s, sys: 0.03 s, total: 1.61 s
       9 Wall time: 1.63 #ring with 2000 variables
      10 sage: time for i in range(1000): _ =x1*y1
      11 CPU times: user 0.21 s, sys: 0.03 s, total: 0.24 s
      12 Wall time: 0.24 #ring with two variables
    
    This is due to the way multivariate polynomials in SAGE are represented. I want to come up with a more sparse representation which does not add zero to zero 1998 times in the second example. (I rewrote the polynomial representation to use dicts of dicts which map indices to exponents e.g., {{1:2}:3} represents 3*y^2 if y is the second variable in the ring.)

Other stuff

I'm a computer science grad student from Bremen, Germany, with a strong interest in cryptanalysis, right now mainly algebraic attacks on block ciphers. I maintain a blog at http://www.informatik.uni-bremen.de/~malb/blog.php .


CategoryHomepage

MartinAlbrecht (last edited 2011-11-10 13:47:27 by malb)