Batch version of the back-propagation algorithm. % Given a set of corresponding input-output pairs and an initial network % [W1,W2,critvec,iter]=batbp(NetDef,W1,W2,PHI,Y,trparms) trains the % network with backpropagation. % % The activation functions must be either linear or tanh. The network % architecture is defined by the matrix NetDef consisting of two % rows. The first row specifies the hidden layer while the second % specifies the output layer. %
标签: back-propagation corresponding input-output algorithm
上传时间: 2016-12-27
上传用户:exxxds
% Train a two layer neural network with the Levenberg-Marquardt % method. % % If desired, it is possible to use regularization by % weight decay. Also pruned (ie. not fully connected) networks can % be trained. % % Given a set of corresponding input-output pairs and an initial % network, % [W1,W2,critvec,iteration,lambda]=marq(NetDef,W1,W2,PHI,Y,trparms) % trains the network with the Levenberg-Marquardt method. % % The activation functions can be either linear or tanh. The % network architecture is defined by the matrix NetDef which % has two rows. The first row specifies the hidden layer and the % second row specifies the output layer.
标签: Levenberg-Marquardt desired network neural
上传时间: 2016-12-27
上传用户:jcljkh
Train a two layer neural network with a recursive prediction error % algorithm ("recursive Gauss-Newton"). Also pruned (i.e., not fully % connected) networks can be trained. % % The activation functions can either be linear or tanh. The network % architecture is defined by the matrix NetDef , which has of two % rows. The first row specifies the hidden layer while the second % specifies the output layer.
标签: recursive prediction algorithm Gauss-Ne
上传时间: 2016-12-27
上传用户:ljt101007
这是LLE的原始算法,原文的参考文献是:S.T.Roweis and L.K.Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290, 2000.
上传时间: 2013-12-20
上传用户:虫虫虫虫虫虫
数值分析高斯——列主元消去法主程序 说明如下: % a----input,matrix of coefficient % b----input,right vector % sol----output,returns the solution of linear equation
标签: input coefficient matrix vector
上传时间: 2017-01-01
上传用户:dancnc
计算全息close all clc clear A=zeros(64) A(15:20,20:40)=1 A(15:50,20:25)=1 A(45:50,20:40)=1 A(30:34,20:35)=1 % ppp=exp(rand(64)*pi*2*i) A=A.*ppp % Author s email: zjliu2001@163.com figure imshow(abs(A),[]) Fa=fft2(fftshift(A)) Fs=fftshift(Fa) Am=abs(Fs) % amplitude Ph=angle(Fs) % phase s=11 % 这表示边长吗? cgh=zeros(64*s) th=max(max(abs(Fs)))
上传时间: 2014-10-13
上传用户:wweqas
密码学界牛人Victor Shoup用C++编写数论类库。 NTL is a high-performance, portable C++ library providing data structures and algorithms for arbitrary length integers for vectors, matrices, and polynomials over the integers and over finite fields and for arbitrary precision floating point arithmetic. NTL provides high quality implementations of state-of-the-art algorithms for: * arbitrary length integer arithmetic and arbitrary precision floating point arithmetic * polynomial arithmetic over the integers and finite fields including basic arithmetic, polynomial factorization, irreducibility testing, computation of minimal polynomials, traces, norms, and more * lattice basis reduction, including very robust and fast implementations of Schnorr-Euchner, block Korkin-Zolotarev reduction, and the new Schnorr-Horner pruning heuristic for block Korkin-Zolotarev * basic linear algebra over the integers, finite fields, and arbitrary precision floating point numbers.
标签: high-performance providing portable library
上传时间: 2014-01-04
上传用户:exxxds
SuperLU is a general purpose library for the direct solution of large, sparse, nonsymmetric systems of linear equations on high performance machines. The library is written in C and is callable from either C or Fortran. The library routines will perform an LU decomposition with partial pivoting and triangular system solves through forward and back substitution. The LU factorization routines can handle non-square matrices but the triangular solves are performed only for square matrices. The matrix columns may be preordered (before factorization) either through library or user supplied routines. This preordering for sparsity is completely separate from the factorization. Working precision iterative refinement subroutines are provided for improved backward stability. Routines are also provided to equilibrate the system, estimate the condition number, calculate the relative backward error, and estimate error bounds for the refined solutions.
标签: nonsymmetric solution SuperLU general
上传时间: 2017-02-20
上传用户:lepoke
cordic methods describe essentially the same algorithm that with suitably chosen inputs can be used to calculate a whole range of scientific functions including sin, cos, tan, arctan, arcsin, arccos, sinh, cosh, tanh, arctanh, log, exp, square root and even multiply and divide. the method dates back to volder [1959], and due to its versatility and compactness, it made possible the microcoding of the hp35 pocket scientific calculator in 1972. here is some code to illustrate the techniques. ive split the methods into three parts linear, circular and hyperbolic. in the hp35 microcode these would be unified into one function (for space reasons). because the linear mode can perform multiply and divide, you only need add/subtract and shift to complete the implementation. you can select in the code whether to do the multiples and divides also by cordic means. other multiplies and divides are all powers of 2 (these dont count). to eliminate these too, would involve ieee hackery.
标签: essentially algorithm describe suitably
上传时间: 2017-03-02
上传用户:litianchu
This is GPS Acquisition..by Matlab, this file performs cold start acquisition on the collected "data". It searches for GPS signals of all satellites, which are listed in field "acqSatelliteList" in the settings structure. Function saves code phase and frequency of the detected signals in the "acqResults" structure.
标签: Acquisition acquisition collected performs
上传时间: 2013-12-28
上传用户:wxhwjf