diff --git a/.gitignore b/.gitignore new file mode 100644 index 0000000..47df1f2 --- /dev/null +++ b/.gitignore @@ -0,0 +1,9 @@ +.DS_Store + +*~ + +*tar.gz + +bin/* +lib/* +obj/* \ No newline at end of file diff --git a/Makefile b/Makefile new file mode 100644 index 0000000..2858d2f --- /dev/null +++ b/Makefile @@ -0,0 +1,492 @@ +#********************************************************** +# +# This software is part of J.-S. Caux's ABACUS++ library. +# +# Copyright (c). +# +#----------------------------------------------------------- +# +# File: Makefile +# +# +#***********************************************************/ + +VERSION = ABACUS++G_8 + +# Base directory, tree of all source files, headers +BASEDIR = #/Users/jcaux/WORK/ABACUS++/ +OBJDIR = $(BASEDIR)obj/ +HEADDIR = $(BASEDIR)include/ +LIBDIR = $(BASEDIR)lib/ +SRCDIR = $(BASEDIR)src/ +EXECSDIR = $(BASEDIR)src/EXECS/ +BINDIR = $(BASEDIR)bin/ + +#COMPILE = g++ -Wall -I$(BASEDIR)include/ -L$(LIBDIR) -O3 -stdlib=libstdc++ +#COMPILE = g++ -Wall -I$(BASEDIR)include/ -L$(LIBDIR) -O3 +COMPILE = g++ -I$(BASEDIR)include/ -L$(LIBDIR) -O3 -w -fopenmp +#COMPILE_MPI = mpicxx -I$(BASEDIR)include/ -L$(LIBDIR) -O3 -stdlib=libstdc++ +COMPILE_MPI = mpicxx -I$(BASEDIR)include/ -L$(LIBDIR) -O3 -w -fopenmp +#COMPILE_OMP = g++ -I$(BASEDIR)include/ -L$(LIBDIR) -O3 -w -fopenmp + +VPATH = $(SRCDIR)BETHE:$(SRCDIR)LIEBLIN:$(SRCDIR)COMBI:$(SRCDIR)EXECS:$(SRCDIR)FITTING:$(SRCDIR)HEIS:$(SRCDIR)INTEG:$(SRCDIR)MATRIX:$(SRCDIR)NRG:$(SRCDIR)ODSLF:$(SRCDIR)SCAN:$(SRCDIR)TBA:$(SRCDIR)UTILS:$(SRCDIR)XXX_h0:$(SRCDIR)XXZ_h0:$(SRCDIR)YOUNG + +vpath %.h $(HEADDIR) + +#Headers_JSC = JSC.h JSC_util.h JSC_Bethe.h JSC_Combi.h JSC_Integ.h JSC_Matrix.h JSC_Scan.h JSC_Vect.h JSC_Young.h +Headers_JSC = JSC.h JSC_util.h JSC_Combi.h JSC_Integ.h JSC_Matrix.h JSC_NRG.h JSC_Spec_Fns.h JSC_Vect.h JSC_Young.h +#Headers_all = $(Headers_JSC) JSC_LiebLin.h JSC_Heis.h JSC_ODSLF.h JSC_Scan.h JSC_XXX_h0.h JSC_XXZ_h0.h +Headers_all = $(Headers_JSC) JSC_LiebLin.h JSC_Heis.h JSC_Scan.h JSC_State_Ensemble.h JSC_XXX_h0.h JSC_XXZ_h0.h + +#Objects_BETHE = $(OBJDIR)Bethe_State.o + +Objects_LIEBLIN = $(OBJDIR)LiebLin_Bethe_State.o $(OBJDIR)LiebLin_Chem_Pot.o $(OBJDIR)LiebLin_Matrix_Element_Contrib.o $(OBJDIR)LiebLin_ln_Overlap.o $(OBJDIR)LiebLin_Sumrules.o $(OBJDIR)LiebLin_State_Ensemble.o \ + $(OBJDIR)LiebLin_Tgt0.o $(OBJDIR)LiebLin_Twisted_lnnorm.o $(OBJDIR)LiebLin_Twisted_ln_Overlap.o $(OBJDIR)LiebLin_Utils.o $(OBJDIR)ln_Density_ME.o $(OBJDIR)ln_Psi_ME.o $(OBJDIR)ln_g2_ME.o + +Objects_HEIS = $(OBJDIR)Heis.o $(OBJDIR)Heis_Chem_Pot.o $(OBJDIR)Heis_Sumrules.o $(OBJDIR)Heis_Matrix_Element_Contrib.o \ + $(OBJDIR)ln_Overlap_XXX.o \ + $(OBJDIR)ln_Sz_ME_XXX.o $(OBJDIR)ln_Sz_ME_XXZ.o $(OBJDIR)ln_Sz_ME_XXZ_gpd.o \ + $(OBJDIR)ln_Smin_ME_XXX.o $(OBJDIR)ln_Smin_ME_XXZ.o $(OBJDIR)ln_Smin_ME_XXZ_gpd.o \ + $(OBJDIR)ln_Szz_ME_XXX.o $(OBJDIR)ln_Smm_ME_XXX.o $(OBJDIR)ln_Szm_p_Smz_ME_XXX.o \ + $(OBJDIR)M_vs_H.o \ + $(OBJDIR)XXX_Bethe_State.o $(OBJDIR)XXZ_Bethe_State.o $(OBJDIR)XXZ_gpd_Bethe_State.o + +Objects_ODSLF = $(OBJDIR)ODSLF.o $(OBJDIR)ODSLF_Chem_Pot.o $(OBJDIR)ODSLF_Sumrules.o $(OBJDIR)ODSLF_XXZ_Bethe_State.o $(OBJDIR)ODSLF_Matrix_Element_Contrib.o \ + $(OBJDIR)ln_Sz_ME_ODSLF_XXZ.o $(OBJDIR)ln_Smin_ME_ODSLF_XXZ.o + +Objects_COMBI = $(OBJDIR)Combinatorics.o + +Objects_FITTING = $(OBJDIR)covsrt.o $(OBJDIR)lin_reg.o $(OBJDIR)mrq.o $(OBJDIR)polint.o $(OBJDIR)polint_cx.o + +Objects_INTEG = $(OBJDIR)Integration.o + +Objects_MATRIX = $(OBJDIR)balanc.o $(OBJDIR)det_LU.o $(OBJDIR)det_LU_CX.o $(OBJDIR)eigsrt.o $(OBJDIR)elmhes.o $(OBJDIR)gaussj.o $(OBJDIR)hqr.o \ + $(OBJDIR)jacobi.o $(OBJDIR)lndet_LU.o $(OBJDIR)lndet_LU_dstry.o $(OBJDIR)lndet_LU_CX.o $(OBJDIR)lndet_LU_CX_dstry.o $(OBJDIR)lubksb.o \ + $(OBJDIR)lubksb_CX.o $(OBJDIR)ludcmp.o $(OBJDIR)ludcmp_CX.o $(OBJDIR)pythag.o $(OBJDIR)tqli.o $(OBJDIR)tred2.o + +Objects_NRG = $(OBJDIR)NRG_State_Selector.o $(OBJDIR)NRG_DME_Matrix_Block_builder.o $(OBJDIR)NRG_K_Weight_integrand.o + +#Objects_SCAN = $(OBJDIR)Descendents.o $(OBJDIR)General_Scan.o $(OBJDIR)General_Scan_Parallel.o $(OBJDIR)Particle_Hole_Excitation_Cost.o $(OBJDIR)Scan_Info.o $(OBJDIR)Scan_Thread_List.o +Objects_SCAN = $(OBJDIR)Descendents.o $(OBJDIR)General_Scan.o $(OBJDIR)General_Scan_Parallel.o $(OBJDIR)Particle_Hole_Excitation_Cost.o $(OBJDIR)Scan_Info.o $(OBJDIR)Scan_Thread_Data.o + +Objects_TBA = $(OBJDIR)Root_Density.o $(OBJDIR)TBA_LiebLin.o $(OBJDIR)TBA_XXZ.o $(OBJDIR)TBA_2CBG.o + +#Objects_UTILS = $(OBJDIR)Data_File_Name.o $(OBJDIR)K_and_Omega_Files.o $(OBJDIR)Smoothen_RAW_into_SF.o $(OBJDIR)Smoothen_RAW_into_SF_LiebLin_Scaled.o $(OBJDIR)Sort_RAW_File.o $(OBJDIR)State_Label.o +Objects_UTILS = $(OBJDIR)Data_File_Name.o $(OBJDIR)K_and_Omega_Files.o $(OBJDIR)Smoothen_RAW_into_SF.o $(OBJDIR)Smoothen_RAW_into_SF_LiebLin_Scaled.o $(OBJDIR)Sort_RAW_File.o $(OBJDIR)State_Label.o + +Objects_XXX_h0 = $(OBJDIR)XXX_h0.o + +Objects_XXZ_h0 = $(OBJDIR)XXZ_h0.o + +Objects_YOUNG = $(OBJDIR)Young_Tableau.o + +#Objects_ALL = $(Objects_LIEBLIN) $(Objects_HEIS) $(Objects_COMBI) $(Objects_FITTING) $(Objects_INTEG) $(Objects_MATRIX) $(Objects_NRG) $(Objects_ODSLF) $(Objects_SCAN) $(Objects_TBA) $(Objects_UTILS) $(Objects_XXX_h0) $(Objects_XXZ_h0) $(Objects_YOUNG) +Objects_ALL = $(Objects_LIEBLIN) $(Objects_HEIS) $(Objects_COMBI) $(Objects_FITTING) $(Objects_INTEG) $(Objects_MATRIX) $(Objects_NRG) $(Objects_SCAN) $(Objects_TBA) $(Objects_UTILS) $(Objects_XXX_h0) $(Objects_XXZ_h0) $(Objects_YOUNG) + +#EXECS = $(BINDIR)LiebLin_DSF $(BINDIR)Smoothen_LiebLin_DSF $(BINDIR)Heis_DSF $(BINDIR)Smoothen_Heis_DSF $(BINDIR)Smoothen_Heis_ASF $(BINDIR)Check_RAW_File +#EXECS = $(BINDIR)LiebLin_DSF $(BINDIR)LiebLin_DSF_tester $(BINDIR)Smoothen_LiebLin_DSF $(BINDIR)Smoothen_LiebLin_DSF_DiK $(BINDIR)Heis_DSF $(BINDIR)Smoothen_Heis_DSF $(BINDIR)Smoothen_Heis_ASF $(BINDIR)Check_RAW_File +EXECS = $(BINDIR)LiebLin_DSF $(BINDIR)LiebLin_Data_Daemon $(BINDIR)LiebLin_RAW_File_Stats $(BINDIR)LiebLin_DSF_tester $(BINDIR)LiebLin_DSF_tester_Ix2 $(BINDIR)LiebLin_DSF_MosesState $(BINDIR)LiebLin_DSF_over_Ensemble $(BINDIR)Smoothen_LiebLin_DSF $(BINDIR)Heis_DSF $(BINDIR)Heis_DSF_GeneralState $(BINDIR)Smoothen_Heis_DSF $(BINDIR)Check_RAW_File + + +########################################## +# The library and the set of executables in bin/ are the ultimate targets + +lib$(VERSION).a : $(Objects_ALL) + ar -cru lib$(VERSION).a $(Objects_ALL) + mv lib$(VERSION).a $(BASEDIR)lib/ + $(COMPILE) $(EXECSDIR)LiebLin_DSF.cc -o $(BINDIR)LiebLin_DSF -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_Data_Daemon.cc -o $(BINDIR)LiebLin_Data_Daemon -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_Data_Daemon_Nscaling.cc -o $(BINDIR)LiebLin_Data_Daemon_Nscaling -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_Catalogue_Fixed_c_k_Nscaling.cc -o $(BINDIR)LiebLin_Catalogue_Fixed_c_k_Nscaling -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_RAW_File_Stats.cc -o $(BINDIR)LiebLin_RAW_File_Stats -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_DSF_GeneralState.cc -o $(BINDIR)LiebLin_DSF_GeneralState -l$(VERSION) +# $(COMPILE) $(EXECSDIR)LiebLin_DSF_MosesState.cc -o $(BINDIR)LiebLin_DSF_MosesState -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_DSF_over_Ensemble.cc -o $(BINDIR)LiebLin_DSF_over_Ensemble -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_DSF_tester.cc -o $(BINDIR)LiebLin_DSF_tester -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_DSF_tester_Ix2.cc -o $(BINDIR)LiebLin_DSF_tester_Ix2 -l$(VERSION) +# $(COMPILE) $(EXECSDIR)LiebLin_Moses_tester.cc -o $(BINDIR)LiebLin_Moses_tester -l$(VERSION) + $(COMPILE) $(EXECSDIR)Smoothen_LiebLin_DSF.cc -o $(BINDIR)Smoothen_LiebLin_DSF -l$(VERSION) + $(COMPILE) $(EXECSDIR)Smoothen_LiebLin_DSF_Scaled.cc -o $(BINDIR)Smoothen_LiebLin_DSF_Scaled -l$(VERSION) + $(COMPILE) $(EXECSDIR)Smoothen_LiebLin_DSF_over_Ensemble.cc -o $(BINDIR)Smoothen_LiebLin_DSF_over_Ensemble -l$(VERSION) + $(COMPILE) $(EXECSDIR)Smoothen_LiebLin_DSF_GeneralState.cc -o $(BINDIR)Smoothen_LiebLin_DSF_GeneralState -l$(VERSION) +# $(COMPILE) $(EXECSDIR)Smoothen_LiebLin_DSF_MosesState.cc -o $(BINDIR)Smoothen_LiebLin_DSF_MosesState -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_TBA.cc -o $(BINDIR)LiebLin_TBA -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_TBA_fixed_nbar.cc -o $(BINDIR)LiebLin_TBA_fixed_nbar -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_TBA_fixed_nbar_ebar.cc -o $(BINDIR)LiebLin_TBA_fixed_nbar_ebar -l$(VERSION) +# $(COMPILE) $(EXECSDIR)Smoothen_LiebLin_DSF_DiK.cc -o $(BINDIR)Smoothen_LiebLin_DSF_DiK -l$(VERSION) +# $(COMPILE) $(EXECSDIR)Smoothen_LiebLin_ASF.cc -o $(BINDIR)Smoothen_LiebLin_ASF -l$(VERSION) + $(COMPILE) $(EXECSDIR)Heis_DSF.cc -o $(BINDIR)Heis_DSF -l$(VERSION) + $(COMPILE) $(EXECSDIR)Heis_DSF_GeneralState.cc -o $(BINDIR)Heis_DSF_GeneralState -l$(VERSION) + $(COMPILE) $(EXECSDIR)Smoothen_Heis_DSF.cc -o $(BINDIR)Smoothen_Heis_DSF -l$(VERSION) +# $(COMPILE) $(EXECSDIR)Smoothen_Heis_ASF.cc -o $(BINDIR)Smoothen_Heis_ASF -l$(VERSION) + $(COMPILE) $(EXECSDIR)XXZ_gpd_StagSz_h0.cc -o $(BINDIR)XXZ_gpd_StagSz_h0 -l$(VERSION) +# $(COMPILE) $(EXECSDIR)ODSLF_DSF.cc -o $(BINDIR)ODSLF_DSF -l$(VERSION) +# $(COMPILE) $(EXECSDIR)Smoothen_ODSLF_DSF.cc -o $(BINDIR)Smoothen_ODSLF_DSF -l$(VERSION) + $(COMPILE) $(EXECSDIR)Check_RAW_File.cc -o $(BINDIR)Check_RAW_File -l$(VERSION) + $(COMPILE) $(EXECSDIR)Analyze_RAW_File.cc -o $(BINDIR)Analyze_RAW_File -l$(VERSION) + $(COMPILE) $(EXECSDIR)RAW_File_Stats.cc -o $(BINDIR)RAW_File_Stats -l$(VERSION) + $(COMPILE) $(EXECSDIR)Produce_Sorted_RAW_File.cc -o $(BINDIR)Produce_Sorted_RAW_File -l$(VERSION) + $(COMPILE) $(EXECSDIR)2CBG_ThLim.cc -o $(BINDIR)2CBG_ThLim -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_Fourier_to_x_equal_t.cc -o $(BINDIR)LiebLin_Fourier_to_x_equal_t -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_Fourier_to_t_equal_x.cc -o $(BINDIR)LiebLin_Fourier_to_t_equal_x -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_Fourier_to_x_equal_t_from_RAW.cc -o $(BINDIR)LiebLin_Fourier_to_x_equal_t_from_RAW -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_Fourier_to_t_equal_x_from_RAW.cc -o $(BINDIR)LiebLin_Fourier_to_t_equal_x_from_RAW -l$(VERSION) + $(COMPILE) $(EXECSDIR)LiebLin_Fourier_ssf_to_Qsqx.cc -o $(BINDIR)LiebLin_Fourier_ssf_to_Qsqx -l$(VERSION) + +########################################### +# Interacting LiebLin gas + +$(OBJDIR)LiebLin_Bethe_State.o : LiebLin_Bethe_State.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)LiebLin_Chem_Pot.o : LiebLin_Chem_Pot.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)LiebLin_Matrix_Element_Contrib.o : LiebLin_Matrix_Element_Contrib.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)LiebLin_ln_Overlap.o : LiebLin_ln_Overlap.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)LiebLin_Sumrules.o : LiebLin_Sumrules.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)LiebLin_State_Ensemble.o : LiebLin_State_Ensemble.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)LiebLin_Tgt0.o : LiebLin_Tgt0.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)LiebLin_Twisted_lnnorm.o : LiebLin_Twisted_lnnorm.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)LiebLin_Utils.o : LiebLin_Utils.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Density_ME.o : ln_Density_ME.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)LiebLin_Twisted_ln_Overlap.o : LiebLin_Twisted_ln_Overlap.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Psi_ME.o : ln_Psi_ME.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_g2_ME.o : ln_g2_ME.cc $(Headers_JSC) JSC_LiebLin.h + $(COMPILE) -c $< -o $@ + + +########################################### +# Heisenberg chains + +$(OBJDIR)Heis.o : Heis.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)Heis_Chem_Pot.o : Heis_Chem_Pot.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)Heis_Sumrules.o : Heis_Sumrules.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)Heis_Matrix_Element_Contrib.o : Heis_Matrix_Element_Contrib.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Overlap_XXX.o : ln_Overlap_XXX.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Sz_ME_XXX.o : ln_Sz_ME_XXX.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Szz_ME_XXX.o : ln_Szz_ME_XXX.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Smm_ME_XXX.o : ln_Smm_ME_XXX.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Szm_p_Smz_ME_XXX.o : ln_Szm_p_Smz_ME_XXX.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Sz_ME_XXZ.o : ln_Sz_ME_XXZ.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Sz_ME_XXZ_gpd.o : ln_Sz_ME_XXZ_gpd.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Smin_ME_XXX.o : ln_Smin_ME_XXX.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Smin_ME_XXZ.o : ln_Smin_ME_XXZ.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Smin_ME_XXZ_gpd.o : ln_Smin_ME_XXZ_gpd.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)M_vs_H.o : M_vs_H.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)XXX_Bethe_State.o : XXX_Bethe_State.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)XXZ_Bethe_State.o : XXZ_Bethe_State.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)XXZ_gpd_Bethe_State.o : XXZ_gpd_Bethe_State.cc $(Headers_JSC) JSC_Heis.h + $(COMPILE) -c $< -o $@ + + +########################################### +# One-d spinless fermions + +$(OBJDIR)ODSLF.o : ODSLF.cc $(Headers_JSC) JSC_Heis.h JSC_ODSLF.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ODSLF_Chem_Pot.o : ODSLF_Chem_Pot.cc $(Headers_JSC) JSC_ODSLF.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ODSLF_Sumrules.o : ODSLF_Sumrules.cc $(Headers_JSC) JSC_ODSLF.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ODSLF_XXZ_Bethe_State.o : ODSLF_XXZ_Bethe_State.cc $(Headers_JSC) JSC_Heis.h JSC_ODSLF.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Smin_ME_ODSLF_XXZ.o : ln_Smin_ME_ODSLF_XXZ.cc $(Headers_JSC) JSC_ODSLF.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ln_Sz_ME_ODSLF_XXZ.o : ln_Sz_ME_ODSLF_XXZ.cc $(Headers_JSC) JSC_ODSLF.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ODSLF_Matrix_Element_Contrib.o : ODSLF_Matrix_Element_Contrib.cc $(Headers_JSC) JSC_ODSLF.h + $(COMPILE) -c $< -o $@ + + +########################################### +# Combinatorics functions + +$(OBJDIR)Combinatorics.o : Combinatorics.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +########################################### +# Integ functions + +$(OBJDIR)Integration.o : Integration.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +########################################### +# Fitting functions + +$(OBJDIR)covsrt.o : covsrt.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)lin_reg.o : lin_reg.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)mrq.o : mrq.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)polint.o : polint.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)polint_cx.o : polint_cx.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + + +########################################### +# Matrix functions + +$(OBJDIR)balanc.o : balanc.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)det_LU.o : det_LU.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)det_LU_CX.o : det_LU_CX.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)eigsrt.o : eigsrt.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)elmhes.o : elmhes.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)gaussj.o : gaussj.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)hqr.o : hqr.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)jacobi.o : jacobi.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)lndet_LU.o : lndet_LU.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)lndet_LU_CX.o : lndet_LU_CX.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)lndet_LU_dstry.o : lndet_LU_dstry.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)lndet_LU_CX_dstry.o : lndet_LU_CX_dstry.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)lubksb.o : lubksb.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)lubksb_CX.o : lubksb_CX.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ludcmp.o : ludcmp.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)ludcmp_CX.o : ludcmp_CX.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)pythag.o : pythag.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)tqli.o : tqli.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)tred2.o : tred2.cc $(Headers_JSC) + $(COMPILE) -c $< -o $@ + + +########################################### +# NRG + +$(OBJDIR)NRG_K_Weight_integrand.o : NRG_K_Weight_integrand.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)NRG_State_Selector.o : NRG_State_Selector.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)NRG_DME_Matrix_Block_builder.o : NRG_DME_Matrix_Block_builder.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + + +########################################### +# Scan + +#$(OBJDIR)Base.o : Base.cc $(Headers_all) +# $(COMPILE) -c $< -o $@ + +$(OBJDIR)Descendents.o : Descendents.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)General_Scan.o : General_Scan.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)General_Scan_Parallel.o : General_Scan_Parallel.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)Particle_Hole_Excitation_Cost.o : Particle_Hole_Excitation_Cost.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +#$(OBJDIR)Offsets.o : Offsets.cc $(Headers_all) +# $(COMPILE) -c $< -o $@ + +$(OBJDIR)Scan_Info.o : Scan_Info.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +#$(OBJDIR)Scan_State_List.o : Scan_State_List.cc $(Headers_all) +# $(COMPILE) -c $< -o $@ + +#$(OBJDIR)Scan_Thread_List.o : Scan_Thread_List.cc $(Headers_all) +#$(OBJDIR)Scan_Thread_Set.o : Scan_Thread_Set.cc $(Headers_all) +$(OBJDIR)Scan_Thread_Data.o : Scan_Thread_Data.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +########################################### +# Thermodynamic Bethe Ansatz + +$(OBJDIR)Root_Density.o : Root_Density.cc $(Headers_JSC) JSC_TBA.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)TBA_LiebLin.o : TBA_LiebLin.cc $(Headers_JSC) JSC_TBA.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)TBA_XXZ.o : TBA_XXZ.cc $(Headers_JSC) JSC_TBA.h + $(COMPILE) -c $< -o $@ + +$(OBJDIR)TBA_2CBG.o : TBA_2CBG.cc $(Headers_JSC) JSC_TBA.h + $(COMPILE) -c $< -o $@ + +########################################### +# Utilities + +$(OBJDIR)Data_File_Name.o : Data_File_Name.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)K_and_Omega_Files.o : K_and_Omega_Files.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)Smoothen_RAW_into_SF.o : Smoothen_RAW_into_SF.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)Smoothen_RAW_into_SF_LiebLin_Scaled.o : Smoothen_RAW_into_SF_LiebLin_Scaled.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)Sort_RAW_File.o : Sort_RAW_File.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +$(OBJDIR)State_Label.o : State_Label.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + +########################################### +# XXX_h0 + +$(OBJDIR)XXX_h0.o : XXX_h0.cc $(Headers_all) JSC_XXX_h0.h + $(COMPILE) -c $< -o $@ + +########################################### +# XXZ_h0 + +$(OBJDIR)XXZ_h0.o : XXZ_h0.cc $(Headers_all) JSC_XXZ_h0.h + $(COMPILE) -c $< -o $@ + +########################################### +# Young Tableaux + +$(OBJDIR)Young_Tableau.o : Young_Tableau.cc $(Headers_all) + $(COMPILE) -c $< -o $@ + + +############################################ + +#executables : +# $(COMPILE) $(EXECSDIR)LiebLin_DSF.cc -o $(BINDIR)LiebLin_DSF -l$(VERSION) +# $(COMPILE) $(EXECSDIR)Smoothen_LiebLin_DSF.cc -o $(BINDIR)Smoothen_LiebLin_DSF -l$(VERSION) +# $(COMPILE) $(EXECSDIR)Heis_DSF.cc -o $(BINDIR)Heis_DSF -l$(VERSION) +# $(COMPILE) $(EXECSDIR)Smoothen_Heis_DSF.cc -o $(BINDIR)Smoothen_Heis_DSF -l$(VERSION) + +########################################### +# Executables (parallel versions) + +parallel : +# $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_par.cc -o $(BINDIR)LiebLin_DSF_par -l$(VERSION) + $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_par_Prepare.cc -o $(BINDIR)LiebLin_DSF_par_Prepare -l$(VERSION) + $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_par_Run.cc -o $(BINDIR)LiebLin_DSF_par_Run -l$(VERSION) + $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_par_Wrapup.cc -o $(BINDIR)LiebLin_DSF_par_Wrapup -l$(VERSION) + $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_GeneralState_par_Prepare.cc -o $(BINDIR)LiebLin_DSF_GeneralState_par_Prepare -l$(VERSION) + $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_GeneralState_par_Run.cc -o $(BINDIR)LiebLin_DSF_GeneralState_par_Run -l$(VERSION) + $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_GeneralState_par_Wrapup.cc -o $(BINDIR)LiebLin_DSF_GeneralState_par_Wrapup -l$(VERSION) +# $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_MosesState_par.cc -o $(BINDIR)LiebLin_DSF_MosesState_par -l$(VERSION) +# $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_MosesState_par_Prepare.cc -o $(BINDIR)LiebLin_DSF_MosesState_par_Prepare -l$(VERSION) +# $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_MosesState_par_Run.cc -o $(BINDIR)LiebLin_DSF_MosesState_par_Run -l$(VERSION) +# $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_MosesState_par_Wrapup.cc -o $(BINDIR)LiebLin_DSF_MosesState_par_Wrapup -l$(VERSION) +# $(COMPILE_MPI) $(EXECSDIR)LiebLin_DSF_over_Ensemble_par.cc -o $(BINDIR)LiebLin_DSF_over_Ensemble_par -l$(VERSION) +# $(COMPILE_MPI) $(EXECSDIR)Heis_DSF_par.cc -o $(BINDIR)Heis_DSF_par -l$(VERSION) + $(COMPILE_MPI) $(EXECSDIR)Heis_DSF_par_Prepare.cc -o $(BINDIR)Heis_DSF_par_Prepare -l$(VERSION) + $(COMPILE_MPI) $(EXECSDIR)Heis_DSF_par_Run.cc -o $(BINDIR)Heis_DSF_par_Run -l$(VERSION) + $(COMPILE_MPI) $(EXECSDIR)Heis_DSF_par_Wrapup.cc -o $(BINDIR)Heis_DSF_par_Wrapup -l$(VERSION) + + +########################################### +# Cleanup + +clean : + /bin/rm -f $(Objects_ALL) + /bin/rm -f $(LIBDIR)$(VERSION) + /bin/rm -f $(EXECS) + + diff --git a/README.md b/README.md index 6abedd7..04a0740 100644 --- a/README.md +++ b/README.md @@ -1,2 +1,235 @@ # ABACUS +Copyright (c) J.-S. Caux + + + +--- + +Previous versions +================= + +------------------------------------------------------------------ +## ABACUS++G_8 [started 2015 02 20] [published 2015 04 04] + +Building up on ++G_7. + +Objectives: + +- facilitate use of OpenMP, by removing the recursive aspect of descending. DONE. + +- improve memory use. DONE. + Replace in-memory Scan_Thread_Set by on-disk Scan_Thread_Data. + Uses Scan_Thread structure. + +Features: +- OpenMP is implemented. +- Scanning on spin chains is now implemented. +- Scan_Info's Nfull now contains the sub-Hilbert space dimension (for spin chains; type == double) +- Added the .stat file in General_Scan, containing info about expected data value and actually obtained one +- Timing is more strictly enforced, using OpenMP wtime(). + +Notes: +- a first implementation using OpenMP is archived as ABACUS++G_8_v1.tar.gz package. + +------------------------------------------------------------------ +ABACUS++G_7 [published 2015 01 17] + +Building up on ++G_6. +- Now using 15 types of descendents, enabling fixed iK scanning with increasing ph nrs. +- Scan_Thread_List replaced by Scan_Thread_Set (to optimize performance and threads storage) + +Works well for LiebLin, including at finite T. + +TODO: +- implement scanning for spin chains (scanning over bases still missing for generic AveragingState) + +------------------------------------------------------------------ +ABACUS++G_6 [started 2015 01] + +Simple scanning: iK stepped up, iK stepped down (from ++G_5). +Fixed momentum scanning is thus implemented. + +Version 6.0: [published 2015 01 16] +- uses 8 types of descendents +- Works well for LiebLin, including finite T. + +Version 6.1: proto version of ++G_7 +- uses greater nr of descendents, 15, enforcing strict ph nr increase + +------------------------------------------------------------------ +ABACUS++G_5 [started 2014 12 11] [abandoned 2015 01, except for version 5.0 (great for GS of LiebLin)] + +Fundamental rewriting of scanning protocol. + +Version 5.0 works very well for ground state correlations of Lieb-Liniger. +Heis: still bugged. + +------------------------------------------------------------------ +ABACUS++G_4 [started 2014 12 08] [abandoned 2014 12 11] + +Nontrivial bug: descending with inner and outer skeletons does not always +preserve number of p-h excitations. Fatal for large c in LiebLin. + +- Descending with type 2: instead of summing over all new p and h positions, + just put p and h as close to each other as possible. + To implement this, added a new type of desc (4), where this distance is increased. + +TODO: +- For XXX: inclusion of infinite rapidities as genuine particle type + + +------------------------------------------------------------------ +ABACUS++G_3 [started 2014 11 10] [closed 2014 12 08] + +- Change of naming convention: LiebLin instead of Bose or 1DBG. +- For LiebLin: changed rapidity naming convention: now lambdaoc, to make rescaling \lambda/c explicit +- Improved small c algorithm for LiebLin +- Fixed momentum scanning implemented using inner and outer skeleton logic + +------------------------------------------------------------------ +ABACUS++G_2 [closed 2014 11 10] +[2013 09 20] +- Scanning is now automatically over remaining pair of excitations; facilitates fixed iK scans +- Threads are now over states with NScan-2 particles, intermediate states then include all Nscan states fulfilling the momentum constraints +- Scanning for spin chains implemented (general states; also parallel). Needs further testing. Seems to work for ground state. + +------------------------------------------------------------------ +ABACUS++G_1.1 + +[2013 09 06] +Changed parallel version for Bose: +- supercycle time is now an argument to functions +- split the 'Prepare', 'Run' and 'Wrapup' into separate executables +- added the 'filenamesuffix' argument to parallel functions + + +------------------------------------------------------------------ +ABACUS++G_1 + +New labelling and descending implemented for arbitrary states with strings, e.g. spin chains. + +Descendents for an arbitrary state are now generated according to the following logic. + +TO DO: +Include string deviation value in Bethe_State objects and in RAW files. Form factors are always computed. + + +[2013 08 25] +Changed sum rules for LiebLin density-density: now using f-sumrule from iKmin to iKmax. + + +Earlier notes for ABACUS++T: +############################################################## + +Notable changes from ABACUS++: + +- The state labelling is now done with a single `string' label, detailing the base, nr of particle-hole excitations (as compared to a reference scan state) and Ix2 changes. The number of particle-hole excitations is therefore not limited anymore, and (for LiebLin) there is no upper momentum limit. See the src/State_Label.cc file for the implementation details. + +- The state scanning procedure (i.e. the descendents logic) is now implemented solely at the level of the quantum numbers, following a recursive logic in which excitations are created at the right, then left Fermi boundary and allowed to propagate deeper outside/inside as the tree is climbed. + +- Since the labelling and descending are now completely general, dynamical correlations not only on ground states but also on arbitrary excited states can be computed efficiently. This thus allows to deal with finite temperature correlations. + + +############################################################## + +IMPROVEMENTS YET TO BE IMPLEMENTED: + +- Some form of binning to make very large runs possible. +Possibility: post-processing of the .raw file, joining together contributions at given iK and (about) same energy into effective ones and saving those into a pre-defined matrix format. +Difficulties: the size of the .raw file is of the order of the .thr one, which must be kept. + +############################################################## + +VERSIONS: NOTES + +Important conventions: + +- Versions are numbered with two integers, [VERSION].[SUBVERSION] +- Changes in subversion number indicate `internal' revisions which do not change any of the external conventions, data file structures, etc. Data produced with version V.S can be `polished' with version V.S' with S' != S. +- Changes in version number indicate a larger scale revision affecting the conventions. Data produced with earlier versions are then deprecated. + +--------------------------------------------------------------- + +ABACUS++T_1.0 [published 23 June 2011] + +First version of new setup and logic for ABACUS. Implementation of all basic new ideas for state labelling and descending. + + +------------------------------------------------------------------ + +ABACUS++T_2.0 [in progress][*ABANDONED*] + +Attempt at implementing fixed-momentum-based scanning. + + +------------------------------------------------------------------ + +ABACUS++T_3.0 [published 1 Nov 2011] + +New, simpler logic for descendents: the hole positions are scanned immediately +upon creation of a new particle-hole pair; the hole positions are then kept fixed +in subsequent descending. + +Problems: +- when scanning for a function such as the one-body function for Lieb-Liniger, +the extra particle in the AveragingState is not scanned to. Version 4 will +address this problem. + +------------------------------------------------------------------ + +ABACUS++T_4.0 [published 1 Nov 2011] + +Modification to the labeling logic: +states are now always labeled using the AveragingState's quantum numbers, +even if their bases are different. The form of the State_Label is thus changed. + +------------------------------------------------------------------ + +ABACUS++T_5.0 [published 4 Nov 2011] + +Introduced two types of threads, separating scanning over fixed particle-hole numbers +and adding a p-h. + +------------------------------------------------------------------ + +ABACUS++T_6.0 [published 21 Nov 2011] + +The scanning is now also done over the hole positions. +Upon the creation of a particle-hole, only hole positions at the edges of +blocks of contiguous Ix2 in OriginState are used. +The holes are then scanned moving towards the center of the blocks. + +There are now thus 3 types of scanning: +0: over holes +1: over particles +2: adding a p-h pair + +------------------------------------------------------------------ + +ABACUS++T_7.0 [published 1 Dec 2011]] + +Optimization of runtime memory use and of output files sizes. + +Labels: introduction of compressed labels (used in raw and thr files). +Removed conv boolean from raw file (all states in this file have converged). +Threads: removed omega and iK (changed General_Scan accordingly), use compressed labels. + +Scanning function: it's now possible to give a default file name when invoking +General_Scan (and thus Scan_Bose, etc). This is done to avoid stupidly long file +names when calculating correlators over AveragingStates which are far from +the ground state. + +------------------------------------------------------------------ + +ABACUS++T_8.0 + +Changed labels slightly, to avoid identifying empty string with integer 0. +All labels with at least one excitation are now strictly nonempty. + + +------------------------------------------------------------------ + +ABACUS++T_9 [in development][abandoned] + +Scanning over ensembles. diff --git a/include/JSC.h b/include/JSC.h new file mode 100644 index 0000000..b0def99 --- /dev/null +++ b/include/JSC.h @@ -0,0 +1,122 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC.h + +Purpose: Core header file, includes all descendents. + + +***********************************************************/ + +#ifndef _JSC_H_ +#define _JSC_H_ + +// This core header file includes all the others + +const char JSC_VERSION[20] = "ABACUS++G_8"; + +// Standard includes +#include +#include // for complex number algebra +#include + +#include +#include +#include + +#include // for exit(), ... +#include // for clock(), ... +#include // for numeric limits +#include // for numeric limits + +// For file sizes, etc +#include "sys/types.h" +#include "sys/stat.h" +#include "unistd.h" + +// Signal handling +#include /* standard I/O functions */ +#include /* standard unix functions, like getpid() */ +#include /* various type definitions, like pid_t */ +#include /* signal name macros, and the signal() prototype */ + +// My own math functions and shortcuts +#include "JSC_util.h" + +// Vectors and matrices +#include "JSC_Vect.h" // My vector class definitions +#include "JSC_Matrix.h" // My matrix class definitions + +// Choose_Table +#include "JSC_Combi.h" + +// Fitting, interpolating +#include "JSC_Fitting.h" + +// Young tableaux +#include "JSC_Young.h" + +// Integration +#include "JSC_Integ.h" + +// Special functions: +#include "JSC_Spec_Fns.h" + +//*** Integrable models: + +// Heisenberg spin-1/2 antiferromagnet +#include "JSC_Heis.h" + +// Lieb-Liniger +#include "JSC_LiebLin.h" + +// One-d spinless fermions: +#include "JSC_ODSLF.h" + +// General: +//#include "JSC_Bethe.h" + +// Thermodynamic Bethe Ansatz utilities +#include "JSC_TBA.h" + +// State ensembles +#include "JSC_State_Ensemble.h" + +// XXX in zero field: Uq(sl(2)) stuff +#include "JSC_XXX_h0.h" + +// XXZ in zero field: quantum groups +#include "JSC_XXZ_h0.h" + +// Two-component Bose gas +//#include "2CBG.h" + +// Richardson +//#include "Richardson.h" + +// *** Correlation functions: + +// New scanning protocols for ABACUS++ +#include "JSC_Scan.h" + +// Functions for everybody +//#include "JSC_fns.h" // KEEP THIS INCLUDE LAST, SINCE IT USES PREVIOUS DECLARATIONS + +// Numerical RG: +#include "JSC_NRG.h" + +// OpenMP +#include + +// Typedefs: +typedef double DP; + + +#endif // _JSC_H_ + + diff --git a/include/JSC_Bethe.h b/include/JSC_Bethe.h new file mode 100644 index 0000000..dcc1bce --- /dev/null +++ b/include/JSC_Bethe.h @@ -0,0 +1,109 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_Bethe.h + +Purpose: Declares Bethe state-related classes and functions. + +***********************************************************/ + +#ifndef _BETHE_ +#define _BETHE_ + +#include "JSC.h" + +namespace JSC { + + /* + class Model { + + protected: + string Name; // LiebLin, XXZ, XXX, XXZ_gpd + int Nsites; // + int L; + DP interaction; + + public: + Model (string Name_ref, + }; + */ + + + class Base { + + public: + int charge; // charge of the base: number of particles (LiebLin gas), number of down spins (Heis), ... + Vect Nrap; // Nrap[i] contains the number of rapidities of type i, i = 0, ..., Nstrings - 1. + int Nraptot; // total number of strings in this state + Vect Ix2_infty; // Ix2_infty[i] contains the max of BAE function for the (half-)integer I[i], i = 0, Nstrings - 1. + Vect Ix2_min; + Vect Ix2_max; // Ix2_max[i] contains the integer part of 2*I_infty, with correct parity for base. + string baselabel; + + public: + Base (); + Base (const Base& RefBase); // copy constructor + + public: // LiebLin constructors + Base (int N); // Constructor for repulsive LiebLin gas case: one type of particle + + public: // HEIS constructors + //Base (const Heis_Chain& RefChain, int M); // constructs configuration with all Mdown in one-string of +1 parity // DEPRECATED + Base (const Heis_Chain& RefChain, const Vect& Nrapidities); // sets to Nrapidities vector, and checks consistency + Base (const Heis_Chain& RefChain, string baselabel_ref); + + public: // operators + Base& operator= (const Base& RefBase); + bool operator== (const Base& RefBase); + bool operator!= (const Base& RefBase); + + public: // member functions + void Compute_Ix2_limits(const Heis_Chain& RefChain); // computes the Ix2_infty and Ix2_max + }; + + + //**************************************************************************** + + // Class Bethe_State is a wrapper class for all types of eigenstates + + class Bethe_State { + + public: // configuration + Base base; + Vect > Ix2; + Vect > lambda; + + public: // properties + DP E; + int iK; + DP K; + DP lnnorm; + + public: // identification + string label; // this is the relative label by default + Vect > OriginIx2; // to define the relative label + + public: // convergence + DP diffsq; + int conv; + int iter; + int iter_Newton; + + //public: // for descendents, etc + + public: // constructors + Bethe_State(); + Bethe_State (const Bethe_State& RefState); // copy constructor + Bethe_State (const Bethe_State& OriginState, string label_ref); + + LiebLin_Bethe_State& operator= (const LiebLin_Bethe_State& RefState); + + +} // namespace JSC + +#endif diff --git a/include/JSC_Combi.h b/include/JSC_Combi.h new file mode 100644 index 0000000..d7eb31e --- /dev/null +++ b/include/JSC_Combi.h @@ -0,0 +1,51 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) 2006-9. + +----------------------------------------------------------- + +File: JSC_Combi.h + +Purpose: Declares combinatorics-related classes and functions. + +Last modified: 08/10/2009 + +***********************************************************/ + +#ifndef _COMBI_ +#define _COMBI_ + +#include "JSC.h" + +namespace JSC { + + //*********************************************************************** + + class Choose_Table { + + private: + int Npower; + int Npowerp1; + int dim; + unsigned long long int* table; + void Fill_table (); + + public: + Choose_Table (); + Choose_Table (int Npower_ref); + Choose_Table (const Choose_Table& Ref_Choose_Table); // constructs a new object from an existing one + int power(); // returns Npower + unsigned long long int choose (int N, int M); + ~Choose_Table (); + + }; + + + std::ostream& operator<< (std::ostream& s, Choose_Table& Ref_table); + + +} // namespace JSC + +#endif diff --git a/include/JSC_Fitting.h b/include/JSC_Fitting.h new file mode 100644 index 0000000..a10b38f --- /dev/null +++ b/include/JSC_Fitting.h @@ -0,0 +1,37 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC.h + +Purpose: Core header file, includes all descendents. + +***********************************************************/ + +#ifndef _FITTING_ +#define _FITTING_ + +namespace JSC { + + // Functions in src/FITTING directory + void covsrt (SQMat_DP& covar, Vect& ia, const int mfit); + void lin_reg (Vect_DP x, Vect_DP y, Vect_DP sigma, DP& a, DP& b, DP& chisq); + void lin_reg (Vect_DP x, Vect_DP y, DP& a, DP& b, DP& chisq); + void mrqmin (Vect_DP& x, Vect_DP& y, Vect_DP& sig, Vect_DP& a, + Vect& ia, SQMat_DP& covar, SQMat_DP& alpha, DP& chisq, + void funcs(const DP, Vect_DP&, DP&, Vect_DP&), DP& alambda); + void mrqcof (Vect_DP& x, Vect_DP& y, Vect_DP& sig, Vect_DP& a, + Vect& ia, SQMat_DP& alpha, Vect_DP& beta, DP& chisq, + void funcs (const DP, Vect_DP&, DP&, Vect_DP&)); + + // For interpolating: + void polint(Vect_DP& xa, Vect_DP& ya, const DP x, DP& y, DP& dy); + void polint(Vect_CX& xa, Vect_CX& ya, const complex x, complex& y, complex& dy); + +} // namespace JSC + +#endif diff --git a/include/JSC_Heis.h b/include/JSC_Heis.h new file mode 100644 index 0000000..15c604f --- /dev/null +++ b/include/JSC_Heis.h @@ -0,0 +1,777 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Heis.h + +Purpose: Declares Heisenberg chain classes and functions. + + + +***********************************************************/ + +#ifndef _HEIS_ +#define _HEIS_ + +#include "JSC.h" + +namespace JSC { + + // First, some global constants... + + const long long int ID_UPPER_LIMIT = 10000000LL; // max size of vectors we can define without seg fault + const int INTERVALS_SIZE = 100000; // size of Scan_Intervals arrays + const int NBASESMAX = 1000; // max number of bases kept + + const DP ITER_REQ_PREC = 100.0 * MACHINE_EPS_SQ; + //const DP ITER_REQ_PREC = MACHINE_EPS_SQ; + + // Cutoffs on particle numbers + //const int NPARTICLES_MAX = 24; + //const int NHOLES_MAX = NPARTICLES_MAX/2; + //const int MAX_RAPS_ABOVE_ZERO = 10; // max nr of rapidities above lowest type + //const int NPARTICLES_MAX = 2; + //const int NHOLES_MAX = 1; + //const int MAX_TYPES_IN_BASE = 4; // maximal number of particle types we allow in bases + const int MAXSTRINGS = 20; // maximal number of particle types we allow in bases + //const int MAXSTRINGS = 2; // maximal number of particle types we allow in bases + //const DP HEIS_deltaprec = 1.0e-6;//1.0e-4; // maximal string deviation allowed // DEPRECATED in ++T_9 + + const int NEXC_MAX_HEIS = 16; // maximal number of excitations (string binding/unbinding, particle-hole) considered + + //*********************************************************************** + + class Heis_Chain { + + public: + DP J; + DP Delta; + DP anis; // acos(Delta) if Delta < 1.0, 0 if Delta == 1.0, acosh(Delta) if Delta > 1.0 + DP hz; + int Nsites; + int Nstrings; // how many possible strings. The following two arrays have Nstrings nonzero elements. + int* Str_L; // vector (length M) containing the allowed string lengths. Elements that are 0 have no meaning. + int* par; // vector (length M) containing the parities of the strings. Elements that are 0 have no meaning. + // Parities are all +1 except for gapless XXZ subcases + DP* si_n_anis_over_2; // for optimization: sin for XXZ, sinh for XXZ_gpd + DP* co_n_anis_over_2; // for optimization + DP* ta_n_anis_over_2; // for optimization + DP prec; // precision required for computations, always put to ITER_REQ_PREC + + public: + Heis_Chain (); + Heis_Chain (DP JJ, DP DD, DP hh, int NN); // contructor: simply initializes + Heis_Chain (const Heis_Chain& RefChain); // copy constructor; + Heis_Chain& operator= (const Heis_Chain& RefChain); + bool operator== (const Heis_Chain& RefChain); + bool operator!= (const Heis_Chain& RefChain); + ~Heis_Chain(); // destructor + + public: + //void Scan_for_Possible_Bases (int Mdown, Vect& possible_base_id, int& nfound, int nexc_max_used, + // int base_level_to_scan, Vect& Nrapidities); + //Vect Possible_Bases (int Mdown); // returns a vector of possible bases + /* Deactivated in ++G_8 + void Scan_for_Possible_Bases (int Mdown, Vect& possible_base_label, int& nfound, int nexc_max_used, + int base_level_to_scan, Vect& Nrapidities); + Vect Possible_Bases (int Mdown); // returns a vector of possible bases + */ + }; + + //**************************************************************************** + + // Objects in class Heis_Base are a checked vector containing the number of rapidities of allowable types for a given state + + class Heis_Base { + + public: + int Mdown; // total number of down spins + Vect Nrap; // Nrap[i] contains the number of rapidities of type i, i = 0, Nstrings - 1. + int Nraptot; // total number of strings in this state + Vect Ix2_infty; // Ix2_infty[i] contains the max of BAE function for the (half-)integer I[i], i = 0, Nstrings - 1. + Vect Ix2_min; + Vect Ix2_max; // Ix2_max[i] contains the integer part of 2*I_infty, with correct parity for base. + //long long int id; // identification number + double dimH; // dimension of sub Hilbert space associated to this base; use double to avoid max int problems. + string baselabel; // base label + + public: + Heis_Base (); + Heis_Base (const Heis_Base& RefBase); // copy constructor + Heis_Base (const Heis_Chain& RefChain, int M); // constructs configuration with all Mdown in one-string of +1 parity + Heis_Base (const Heis_Chain& RefChain, const Vect& Nrapidities); // sets to Nrapidities vector, and checks consistency + //Heis_Base (const Heis_Chain& RefChain, long long int id_ref); + Heis_Base (const Heis_Chain& RefChain, string baselabel_ref); + inline int& operator[] (const int i); + inline const int& operator[] (const int i) const; + Heis_Base& operator= (const Heis_Base& RefBase); + bool operator== (const Heis_Base& RefBase); + bool operator!= (const Heis_Base& RefBase); + + void Compute_Ix2_limits(const Heis_Chain& RefChain); // computes the Ix2_infty and Ix2_max + + //void Scan_for_Possible_Types (Vect& possible_type_id, int& nfound, int base_level, Vect& Nexcitations); + //Vect Possible_Types (); // returns a vector of possible types + + }; + + inline int& Heis_Base::operator[] (const int i) + { + return Nrap[i]; + } + + inline const int& Heis_Base::operator[] (const int i) const + { + return Nrap[i]; + } + + //**************************************************************************** + /* + // Objects in class Ix2_Config carry all the I's of a given state + + class Ix2_Config { + + //private: + public: + int Nstrings; + Vect Nrap; + int Nraptot; + + //int** Ix2; + Vect > Ix2; + + public: + Ix2_Config (); + Ix2_Config (const Heis_Chain& RefChain, int M); // constructor, puts I's to ground state + Ix2_Config (const Heis_Chain& RefChain, const Heis_Base& base); // constructor, putting I's to lowest-energy config + // consistent with Heis_Base configuration for chain RefChain + Ix2_Config& operator= (const Ix2_Config& RefConfig); + //inline int* operator[] (const int i); + inline Vect operator[] (const int i); + //inline const int* operator[] (const int i) const; + inline const Vect operator[] (const int i) const; + //~Ix2_Config(); // not needed, inherited from Vect + string Return_Label (const Ix2_Config& OriginIx2); + }; + + //inline int* Ix2_Config::operator[] (const int i) + inline Vect Ix2_Config::operator[] (const int i) + { + return Ix2[i]; + } + + //inline const int* Ix2_Config::operator[] (const int i) const + inline const Vect Ix2_Config::operator[] (const int i) const + { + return Ix2[i]; + } + + std::ostream& operator<< (std::ostream& s, const Ix2_Config& RefConfig); + */ + //**************************************************************************** + + // Objects in class Lambda carry all rapidities of a state + + class Lambda { + + private: + int Nstrings; + Vect Nrap; + int Nraptot; + DP** lambda; + //Vect > lambda; + + public: + Lambda (); + Lambda (const Heis_Chain& RefChain, int M); // constructor, puts all lambda's to zero + Lambda (const Heis_Chain& RefChain, const Heis_Base& base); // constructor, putting I's to lowest-energy config + // consistent with Heis_Base configuration for chain RefChain + Lambda& operator= (const Lambda& RefConfig); + inline DP* operator[] (const int i); + //inline Vect operator[] (const int i); + inline const DP* operator[] (const int i) const; + //inline const Vect operator[] (const int i) const; + ~Lambda(); + + }; + + inline DP* Lambda::operator[] (const int i) + //inline Vect Lambda::operator[] (const int i) + { + return lambda[i]; + } + + inline const DP* Lambda::operator[] (const int i) const + //inline const Vect Lambda::operator[] (const int i) const + { + return lambda[i]; + } + + + //**************************************************************************** + + // Objects in class Ix2_Offsets carry Young tableau representations of the Ix2 configurations + /* + class Ix2_Offsets { + + public: + Heis_Base base; + Vect Tableau; // vector of pointers to tableaux at each level + long long int type_id; + long long int id; // id number of offset + long long int maxid; // max id number allowable + + public: + Ix2_Offsets (); + Ix2_Offsets (const Ix2_Offsets& RefOffset); // copy constructor + Ix2_Offsets (const Heis_Base& RefBase, long long int req_type_id); + Ix2_Offsets (const Heis_Base& RefBase, Vect nparticles); // sets all tableaux to empty ones, with nparticles[] at each level + Ix2_Offsets& operator= (const Ix2_Offsets& RefOffset); + bool operator<= (const Ix2_Offsets& RefOffsets); + bool operator>= (const Ix2_Offsets& RefOffsets); + + public: + void Set_to_id (long long int idnr); + void Compute_id (); + void Compute_type_id (); + + public: + bool Add_Boxes_From_Lowest (int Nboxes, bool odd_sectors); // adds Nboxes in minimal energy config, all boxes in either even or odd sectors + + }; + + inline long long int Ix2_Offsets_type_id (Vect& nparticles) + { + long long int type_id_here = 0ULL; + + for (int i = 0; i < nparticles.size(); ++i) + type_id_here += nparticles[i] * pow_ulli(10ULL, i); + + return(type_id_here); + } + + long long int Find_idmin (Ix2_Offsets& scan_offsets, int particle_type, int tableau_level, int Row_L_min); + long long int Find_idmax (Ix2_Offsets& scan_offsets, int particle_type, int tableau_level); + */ + //**************************************************************************** + // Objects in class Ix2_Offsets_List carry a vector of used Ix2_Offsets + /* + class Ix2_Offsets_List { + + public: + int ndef; + Vect Offsets; + + public: + Ix2_Offsets_List (); + Ix2_Offsets& Return_Offsets (Heis_Base& RefBase, Vect nparticles); // returns the Ix2_Offsets corresponding to nparticles[]/base + Ix2_Offsets& Return_Offsets (Heis_Base& RefBase, long long int req_type_id); + }; + */ + //**************************************************************************** + + // Objects in class Heis_Bethe_State carry all information about an eigenstate + + // Derived classes include XXZ_Bethe_State, XXX_Bethe_State, XXZ_gpd_Bethe_State + // These contain subclass-specific functions and data. + + class Heis_Bethe_State { + + public: + Heis_Chain chain; + Heis_Base base; + //Ix2_Offsets offsets; + //Ix2_Config Ix2; + Vect > Ix2; + Lambda lambda; + Lambda deviation; // string deviations + Lambda BE; // Bethe equation for relevant rapidity, in the form BE = theta - (1/N)\sum ... - \pi I/N = 0 + DP diffsq; // sum of squares of rapidity differences in last iteration + int conv; // convergence status + DP dev; // sum of absolute values of string deviations + int iter; // number of iterations necessary for convergence + int iter_Newton; // number of iterations necessary for convergence (Newton method) + DP E; // total energy + int iK; // K = 2.0*PI * iK/Nsites + DP K; // total momentum + DP lnnorm; // ln of norm of reduced Gaudin matrix + //long long int base_id; + //long long int type_id; + //long long int id; + //long long int maxid; + //int nparticles; + string label; + + public: + Heis_Bethe_State (); + Heis_Bethe_State (const Heis_Bethe_State& RefState); // copy constructor + //Heis_Bethe_State (const Heis_Bethe_State& RefState, long long int type_id_ref); // new state with requested type_id + Heis_Bethe_State (const Heis_Chain& RefChain, int M); // constructor to ground-state configuration + Heis_Bethe_State (const Heis_Chain& RefChain, const Heis_Base& base); // constructor to lowest-energy config with base + //Heis_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref); + virtual ~Heis_Bethe_State () {}; + + public: + int Charge () { return(base.Mdown); }; + //void Set_Ix2_Offsets (const Ix2_Offsets& RefOffset); // sets the Ix2 to given offsets + //void Set_to_id (long long int id_ref); + //void Set_to_id (long long int id_ref, Heis_Bethe_State& RefState); + //int Nparticles (); // counts the number of particles in state once Ix2 offsets set (so type_id is correctly set) + //void Set_to_Label (string label_ref, const Ix2_Config& OriginIx2); + void Set_to_Label (string label_ref, const Vect >& OriginIx2); + void Set_Label_from_Ix2 (const Vect >& OriginIx2); + bool Check_Symmetry (); // checks whether the I's are symmetrically distributed + void Compute_diffsq (); // \sum BE[j][alpha]^2 + void Find_Rapidities (bool reset_rapidities); // Finds the rapidities + void Find_Rapidities_Twisted (bool reset_rapidities, DP twist); // Finds the rapidities with twist added to RHS of logBE + //void BAE_smackdown (DP max_allowed); + //void Solve_BAE_smackdown (DP max_allowed, int maxruns); + void Solve_BAE_bisect (int j, int alpha, DP req_prec, int itermax); + void Iterate_BAE (DP iter_factor); // Finds new set of lambda[j][alpha] from previous one by simple iteration + void Solve_BAE_straight_iter (DP straight_prec, int max_iter_interp, DP iter_factor); + void Solve_BAE_extrap (DP extrap_prec, int max_iter_extrap, DP iter_factor); + void Iterate_BAE_Newton (); // Finds new set of lambda[j][alpha] from previous one by a Newton step + void Solve_BAE_Newton (DP Newton_prec, int max_iter_Newton); + void Solve_BAE_with_silk_gloves (DP silk_prec, int max_iter_silk, DP iter_factor); + void Compute_lnnorm (); + void Compute_Momentum (); + void Compute_All (bool reset_rapidities); // solves BAE, computes E, K and lnnorm + + inline bool Set_to_Inner_Skeleton (int iKneeded, const Vect >& OriginStateIx2) + { + Ix2[0][0] = Ix2[0][1] - 2; + Ix2[0][base.Nrap[0] - 1] = Ix2[0][base.Nrap[0] - 2] + 2; + (*this).Compute_Momentum(); + if (base.Nrap[0] == 0) return(false); + if (iKneeded >= iK) Ix2[0][base.Nrap[0]-1] += 2*(iKneeded - iK); + else Ix2[0][0] += 2*(iKneeded - iK); + if (Ix2[0][0] < base.Ix2_min[0] || Ix2[0][base.Nrap[0]-1] > base.Ix2_max[0]) return(false); + (*this).Set_Label_from_Ix2 (OriginStateIx2); + return(true); + } + void Set_to_Outer_Skeleton (const Vect >& OriginStateIx2) { + Ix2[0][0] = base.Ix2_min[0] - 4; + Ix2[0][base.Nrap[0]-1] = base.Ix2_max[0] + 4; + (*this).Set_Label_from_Ix2 (OriginStateIx2); + }; + + void Set_to_Closest_Matching_Ix2_fixed_Base (const Heis_Bethe_State& StateToMatch); // defined in Heis.cc + + + // Virtual functions, all defined in the derived classes + + public: + virtual void Set_Free_lambdas() { JSCerror("Heis_Bethe_State::..."); } // sets the rapidities to solutions of BAEs without scattering terms + virtual bool Check_Admissibility(char option) { JSCerror("Heis_Bethe_State::..."); return(false); } + // verifies that we don't have a symmetrical Ix2 config with a Ix2 == 0 for a string of even length >= 2. + virtual void Compute_BE (int j, int alpha) { JSCerror("Heis_Bethe_State::..."); } + virtual void Compute_BE () { JSCerror("Heis_Bethe_State::..."); } + virtual DP Iterate_BAE(int i, int alpha) { JSCerror("Heis_Bethe_State::..."); return(0.0);} + virtual bool Check_Rapidities() { JSCerror("Heis_Bethe_State::..."); return(false); } + virtual DP String_delta () { JSCerror("Heis_Bethe_State::..."); return(0.0); } + virtual void Compute_Energy () { JSCerror("Heis_Bethe_State::..."); } + virtual void Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red) { JSCerror("Heis_Bethe_State::..."); } + }; + + inline bool Is_Inner_Skeleton (Heis_Bethe_State& State) { + return (State.base.Nrap[0] >= 2 && (State.Ix2[0][0] == State.Ix2[0][1] - 2 || State.Ix2[0][State.base.Nrap[0]-1] == State.Ix2[0][State.base.Nrap[0]-2] + 2)); + }; + inline bool Is_Outer_Skeleton (Heis_Bethe_State& State) { + return (State.Ix2[0][0] == State.base.Ix2_min[0] - 4 && State.Ix2[0][State.base.Nrap[0]-1] == State.base.Ix2_max[0] + 4); + }; + + inline bool Force_Descent (char whichDSF, Heis_Bethe_State& ScanState, Heis_Bethe_State& RefState, int desc_type_required, int iKmod, DP Chem_Pot) + { + bool force_descent = false; + + // Force descent if energy of ScanState is lower than that of RefState + //if (ScanState.E - RefState.E - (ScanState.base.Mdown - RefState.base.Mdown) < 0.0) return(true); + /* + // We force descent if + // 1) - there exists a higher string whose quantum number is still on 0 + // AND - there is at most a single particle-hole in the 0 base level + // AND - either the particle or the hole hasn't yet moved. + if (ScanState.base_id/100000LL > 0) { // there is a higher string + int type0 = ScanState.type_id % 10000; + if (type0 == 0 + || type0 == 101 && ScanState.offsets.Tableau[0].id * ScanState.offsets.Tableau[2].id == 0LL + || type0 == 110 && ScanState.offsets.Tableau[1].id * ScanState.offsets.Tableau[2].id == 0LL + || type0 == 1001 && ScanState.offsets.Tableau[0].id * ScanState.offsets.Tableau[3].id == 0LL + || type0 == 1010 && ScanState.offsets.Tableau[1].id * ScanState.offsets.Tableau[3].id == 0LL) // single p-h pair in base level 0 + for (int j = 1; j < ScanState.chain.Nstrings; ++j) { + if (ScanState.base[j] == 1 && ScanState.Ix2[j][0] == 0) { + force_descent = true; + } + } + } + */ + // Force descent if quantum nr distribution is symmetric: + //if (ScanState.Check_Symmetry()) force_descent = true; + //if (desc_type_required > 8 && ScanState.Check_Symmetry()) force_descent = true; + + // Force descent for longitudinal if we're at zero or pi momentum: + //ScanState.Compute_Momentum(); + //if (whichDSF == 'z' && (ScanState.iK - RefState.iK) % iKmod == 0) force_descent = true; + //if (desc_type_required > 8 && whichDSF == 'z' && (2*(ScanState.iK - RefState.iK) % iKmod == 0)) force_descent = true; // type_req > 8 means that we don't conserve momentum + // Force descent for all DSFs if we're at K = 0 or PI and not conserving momentum upon descent: + if (desc_type_required > 8 && (2*(ScanState.iK - RefState.iK) % iKmod == 0)) force_descent = true; // type_req > 8 means that we don't conserve momentum + + //if (force_descent) cout << "Forcing descent on state with label " << ScanState.label << endl; + + if (ScanState.chain.Delta > 1.0) { + /* + // Count the nr of holes in one-strings: + int nholes = 0; + for (int i = 0; i < ScanState.base.Nrap[0] - 1; ++i) if (ScanState.Ix2[0][i+1] - ScanState.Ix2[0][i] != 2) nholes++; + + if (nholes <= 2) { + if (ScanState.base.Nrap[0] == ScanState.base.Mdown - 2 && ScanState.base.Nrap[1] == 1) force_descent = true; + if (ScanState.base.Nrap[0] == ScanState.base.Mdown - 3 && ScanState.base.Nrap[2] == 1) force_descent = true; + if (ScanState.base.Nrap[0] == ScanState.base.Mdown - 4 && ScanState.base.Nrap[1] == 2) force_descent = true; + } + */ + + if (ScanState.label.compare(0, 10, "14x1y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "14x1y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "14x1y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "14x1y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "14x1y1_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "14x1y1_2x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "12x1y2_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "12x1y2_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "12x1y2_0x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "12x1y2_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "12x1y2_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "12x1y2_1x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "12x1y2_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "12x1y2_2x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "12x1y2_2x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "13x2y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "13x2y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "13x2y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "13x2y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "13x2y1_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "13x2y1_2x1") == 0) force_descent = true; + + if (ScanState.label.compare(0, 10, "30x1y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "30x1y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "30x1y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "30x1y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "30x1y1_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "30x1y1_2x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "28x1y2_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "28x1y2_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "28x1y2_0x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "28x1y2_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "28x1y2_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "28x1y2_1x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "28x1y2_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "28x1y2_2x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "28x1y2_2x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "29x2y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "29x2y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "29x2y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "29x2y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "29x2y1_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "29x2y1_2x1") == 0) force_descent = true; + + if (ScanState.label.compare(0, 10, "46x1y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "46x1y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "46x1y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "46x1y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "46x1y1_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "46x1y1_2x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "44x1y2_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "44x1y2_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "44x1y2_0x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "44x1y2_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "44x1y2_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "44x1y2_1x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "44x1y2_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "44x1y2_2x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "44x1y2_2x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "45x2y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "45x2y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "45x2y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "45x2y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "45x2y1_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "45x2y1_2x1") == 0) force_descent = true; + + + if (ScanState.label.compare(0, 10, "62x1y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "62x1y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "62x1y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "62x1y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "62x1y1_2x0") == 0) force_descent = true; + //if (ScanState.label.compare(0, 10, "62x1y1_2x1") == 0 && desc_type_required < 2) force_descent = true; + if (ScanState.label.compare(0, 10, "62x1y1_2x1") == 0 + && (desc_type_required == 14 || desc_type_required == 13 || desc_type_required == 11 || desc_type_required == 10)) force_descent = true; + /* + if (ScanState.label.compare(0, 10, "60x1y2_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "60x1y2_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "60x1y2_0x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "60x1y2_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "60x1y2_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "60x1y2_1x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "60x1y2_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "60x1y2_2x1") == 0 && desc_type_required < 2) force_descent = true; + if (ScanState.label.compare(0, 10, "60x1y2_2x2") == 0 && desc_type_required < 2) force_descent = true; + if (ScanState.label.compare(0, 10, "61x2y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "61x2y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "61x2y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "61x2y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "61x2y1_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "61x2y1_2x1") == 0 && desc_type_required < 2) force_descent = true; + */ + + if (ScanState.label.compare(0, 11, "126x1y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 11, "126x1y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 11, "126x1y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 11, "126x1y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 11, "126x1y1_2x0") == 0) force_descent = true; + //if (ScanState.label.compare(0, 11, "126x1y1_2x1") == 0 && desc_type_required < 2) force_descent = true; + if (ScanState.label.compare(0, 11, "126x1y1_2x1") == 0 + && (desc_type_required == 14 || desc_type_required == 13 || desc_type_required == 11 || desc_type_required == 10)) force_descent = true; + /* + if (ScanState.label.compare(0, 10, "124x1y2_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "124x1y2_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "124x1y2_0x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "124x1y2_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "124x1y2_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "124x1y2_1x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "124x1y2_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "124x1y2_2x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "124x1y2_2x2") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "125x2y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "125x2y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "125x2y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "125x2y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "125x2y1_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 10, "125x2y1_2x1") == 0) force_descent = true; + */ + if (ScanState.label.compare(0, 11, "254x1y1_0x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 11, "254x1y1_0x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 11, "254x1y1_1x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 11, "254x1y1_1x1") == 0) force_descent = true; + if (ScanState.label.compare(0, 11, "254x1y1_2x0") == 0) force_descent = true; + if (ScanState.label.compare(0, 11, "254x1y1_2x1") == 0 && desc_type_required < 2) force_descent = true; + + // Do not force descent a state with rapidities outside of fundamental interval: + /* + for (int j = 0; j < ScanState.chain.Nstrings; ++j) { + // Logic: allow rapidities -PI/2 <= lambda <= PI/2 (including boundaries) + if (ScanState.base.Nrap[j] > 0 && + (ScanState.lambda[j][0] < -PI/2 || ScanState.lambda[j][ScanState.base.Nrap[j] - 1] > PI/2)) + force_descent = false; + + // rapidities should also be ordered as the quantum numbers: + for (int alpha = 1; alpha < ScanState.base.Nrap[j]; ++alpha) + if (ScanState.lambda[j][alpha - 1] >= ScanState.lambda[j][alpha]) + force_descent = false; + } + */ + //if (force_descent) cout << "Forcing descent on state with label " << ScanState.label << endl; + } // if Delta > 1 + + //if (ScanState.base.Nrap[0] == ScanState.base.Mdown - 2 && ScanState.base.Nrap[1] == 1 && ScanState.Ix2[1][0] == 0) force_descent = true; + //if (ScanState.base.Nrap[0] == ScanState.base.Mdown - 3 && ScanState.base.Nrap[2] == 1 && ScanState.Ix2[2][0] == 0) force_descent = true; + //if (ScanState.base.Nrap[0] == ScanState.base.Mdown - 4 && ScanState.base.Nrap[1] == 2 && ScanState.Ix2[1][0] == -1 && ScanState.Ix2[1][1] == 1) force_descent = true; + + return(force_descent); + } + + std::ostream& operator<< (std::ostream& s, const Heis_Bethe_State& state); + + //**************************************************************************** + + // Objects in class XXZ_Bethe_State carry all extra information pertaining to XXZ gapless + + class XXZ_Bethe_State : public Heis_Bethe_State { + + public: + Lambda sinhlambda; + Lambda coshlambda; + Lambda tanhlambda; + + public: + XXZ_Bethe_State (); + XXZ_Bethe_State (const XXZ_Bethe_State& RefState); // copy constructor + XXZ_Bethe_State (const Heis_Chain& RefChain, int M); // constructor to ground-state configuration + XXZ_Bethe_State (const Heis_Chain& RefChain, const Heis_Base& base); // constructor to lowest-energy config with base + //XXZ_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref); // constructor to lowest-energy config with bas + + public: + XXZ_Bethe_State& operator= (const XXZ_Bethe_State& RefState); + + public: + void Set_Free_lambdas(); // sets the rapidities to solutions of BAEs without scattering terms + void Compute_sinhlambda(); + void Compute_coshlambda(); + void Compute_tanhlambda(); + bool Check_Admissibility(char option); // verifies that we don't have a symmetrical Ix2 config with a Ix2 == 0 for a string of even length >= 2. + void Compute_BE (int j, int alpha); + void Compute_BE (); + DP Iterate_BAE(int i, int j); + bool Check_Rapidities(); // checks that all rapidities are not nan + DP String_delta (); + void Compute_Energy (); + //void Compute_Momentum (); + void Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red); + + // XXZ specific functions: + public: + + }; + + XXZ_Bethe_State Add_Particle_at_Center (const XXZ_Bethe_State& RefState); + XXZ_Bethe_State Remove_Particle_at_Center (const XXZ_Bethe_State& RefState); + + //**************************************************************************** + + // Objects in class XXX_Bethe_State carry all extra information pertaining to XXX antiferromagnet + + class XXX_Bethe_State : public Heis_Bethe_State { + + public: + XXX_Bethe_State (); + XXX_Bethe_State (const XXX_Bethe_State& RefState); // copy constructor + XXX_Bethe_State (const Heis_Chain& RefChain, int M); // constructor to ground-state configuration + XXX_Bethe_State (const Heis_Chain& RefChain, const Heis_Base& base); // constructor to lowest-energy config with base + //XXX_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref); // constructor to lowest-energy config with base + + public: + XXX_Bethe_State& operator= (const XXX_Bethe_State& RefState); + + public: + void Set_Free_lambdas(); // sets the rapidities to solutions of BAEs without scattering terms + bool Check_Admissibility(char option); // verifies that we don't have a symmetrical Ix2 config with a Ix2 == 0 for a string of even length >= 2. + void Compute_BE (int j, int alpha); + void Compute_BE (); + DP Iterate_BAE(int i, int j); + bool Check_Rapidities(); // checks that all rapidities are not nan + DP String_delta (); + void Compute_Energy (); + //void Compute_Momentum (); + void Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red); + + // XXX specific functions + public: + bool Check_Finite_rap (); + + }; + + XXX_Bethe_State Add_Particle_at_Center (const XXX_Bethe_State& RefState); + XXX_Bethe_State Remove_Particle_at_Center (const XXX_Bethe_State& RefState); + + //**************************************************************************** + + // Objects in class XXZ_gpd_Bethe_State carry all extra information pertaining to XXZ gapped antiferromagnets + + class XXZ_gpd_Bethe_State : public Heis_Bethe_State { + + public: + Lambda sinlambda; + Lambda coslambda; + Lambda tanlambda; + + public: + XXZ_gpd_Bethe_State (); + XXZ_gpd_Bethe_State (const XXZ_gpd_Bethe_State& RefState); // copy constructor + XXZ_gpd_Bethe_State (const Heis_Chain& RefChain, int M); // constructor to ground-state configuration + XXZ_gpd_Bethe_State (const Heis_Chain& RefChain, const Heis_Base& base); // constructor to lowest-energy config with base + //XXZ_gpd_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref); // constructor to lowest-energy config with base + + public: + XXZ_gpd_Bethe_State& operator= (const XXZ_gpd_Bethe_State& RefState); + + public: + void Set_Free_lambdas(); // sets the rapidities to solutions of BAEs without scattering terms + void Compute_sinlambda(); + void Compute_coslambda(); + void Compute_tanlambda(); + int Weight(); // weight function for contributions cutoff + bool Check_Admissibility(char option); // verifies that we don't have a symmetrical Ix2 config with a Ix2 == 0 for a string of even length >= 2. + void Compute_BE (int j, int alpha); + void Compute_BE (); + DP Iterate_BAE(int i, int j); + void Iterate_BAE_Newton(); + bool Check_Rapidities(); // checks that all rapidities are not nan and are in interval ]-PI/2, PI/2] + DP String_delta (); + void Compute_Energy (); + //void Compute_Momentum (); + void Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red); + + // XXZ_gpd specific functions + public: + + }; + + XXZ_gpd_Bethe_State Add_Particle_at_Center (const XXZ_gpd_Bethe_State& RefState); + XXZ_gpd_Bethe_State Remove_Particle_at_Center (const XXZ_gpd_Bethe_State& RefState); + + //*********************************************** + + // Function declarations + + // in M_vs_H.cc + DP Ezero (DP Delta, int N, int M); + DP H_vs_M (DP Delta, int N, int M); + DP HZmin (DP Delta, int N, int M, Vect_DP& Ezero_ref); + int M_vs_H (DP Delta, int N, DP HZ); + + DP X_avg (char xyorz, DP Delta, int N, int M); + + DP Chemical_Potential (const Heis_Bethe_State& RefState); + DP Particle_Hole_Excitation_Cost (char whichDSF, Heis_Bethe_State& AveragingState); + + //DP Sumrule_Factor (char whichDSF, Heis_Bethe_State& RefState, DP Chem_Pot, bool fixed_iK, int iKneeded); + DP Sumrule_Factor (char whichDSF, Heis_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax); + void Evaluate_F_Sumrule (string prefix, char whichDSF, const Heis_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax); + + complex ln_Sz_ME (XXZ_Bethe_State& A, XXZ_Bethe_State& B); + complex ln_Smin_ME (XXZ_Bethe_State& A, XXZ_Bethe_State& B); + + complex ln_Sz_ME (XXX_Bethe_State& A, XXX_Bethe_State& B); + complex ln_Smin_ME (XXX_Bethe_State& A, XXX_Bethe_State& B); + + // From Antoine Klauser: + complex ln_Szz_ME (XXX_Bethe_State& A, XXX_Bethe_State& B); + complex ln_Szm_p_Smz_ME (XXX_Bethe_State& A, XXX_Bethe_State& B); + complex ln_Smm_ME (XXX_Bethe_State& A, XXX_Bethe_State& B); + + complex ln_Sz_ME (XXZ_gpd_Bethe_State& A, XXZ_gpd_Bethe_State& B); + complex ln_Smin_ME (XXZ_gpd_Bethe_State& A, XXZ_gpd_Bethe_State& B); + + // The following functions have become member functions. + //DP String_delta (XXZ_Bethe_State& state); + //DP String_delta (XXX_Bethe_State& state); + //DP String_delta (XXZ_gpd_Bethe_State& state); + + //DP Compute_Form_Factor_Entry (char whichDSF, Heis_Bethe_State& LeftState, Heis_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile); + //DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXZ_Bethe_State& LeftState, + // XXZ_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile); + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXZ_Bethe_State& LeftState, + XXZ_Bethe_State& RefState, DP Chem_Pot, stringstream& DAT_outfile); + //DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXX_Bethe_State& LeftState, + // XXX_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile); + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXX_Bethe_State& LeftState, + XXX_Bethe_State& RefState, DP Chem_Pot, stringstream& DAT_outfile); + //DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXZ_gpd_Bethe_State& LeftState, + // XXZ_gpd_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile); + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXZ_gpd_Bethe_State& LeftState, + XXZ_gpd_Bethe_State& RefState, DP Chem_Pot, stringstream& DAT_outfile); + + // For geometric quench: + complex ln_Overlap (XXX_Bethe_State& A, XXX_Bethe_State& B); + + void Scan_Heis_Geometric_Quench (DP Delta, int N_1, int M, long long int base_id_1, long long int type_id_1, long long int id_1, + int N_2, int iKmin, int iKmax, int Max_Secs, bool refine); + + +} // namespace JSC + +#endif diff --git a/include/JSC_Infprec.h b/include/JSC_Infprec.h new file mode 100644 index 0000000..e87136f --- /dev/null +++ b/include/JSC_Infprec.h @@ -0,0 +1,32 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_Infprec.h + +Purpose: Declarations for infinite precision arithmetic classes. + +***********************************************************/ + + +#ifndef _JSC_INFPREC_ +#define _JSC_INFPREC_ + +namespace JSC { + + class infprec_int { + + private: + int nintrep; // number of integers used in representation + int baseint; // fundamental int + Vect xint; // extra integers + + } + +} // namespace JSC + +#endif diff --git a/include/JSC_Integ.h b/include/JSC_Integ.h new file mode 100644 index 0000000..a3affe7 --- /dev/null +++ b/include/JSC_Integ.h @@ -0,0 +1,331 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: JSC_Integ.h + +Purpose: Declares combinatorics-related classes and functions. + + +***********************************************************/ + +#ifndef _INTEG_ +#define _INTEG_ + +#include "JSC.h" + +namespace JSC { + + + //********************** Class Domain ************************ + + template + class Domain { + + private: + Vect bdry; + + public: + Domain () : bdry(Vect(2)) + { + bdry[0] = T(0); + bdry[1] = T(0); + } + + public: + Domain (T xmin_ref, T xmax_ref) : bdry(Vect(2)) + { + if (xmax_ref < xmin_ref) JSCerror("Use xmax > xmin in Domain."); + + bdry[0] = xmin_ref; + bdry[1] = xmax_ref; + } + + public: + inline T xmin (int i) + { + if (i > bdry.size()/2) JSCerror("i index too high in Domain::xmin."); + + return(bdry[2*i]); + } + + public: + inline T xmax (int i) + { + if (i > bdry.size()/2) JSCerror("i index too high in Domain::xmax."); + + return(bdry[2*i + 1]); + } + + public: + inline int Ndomains () + { + return(bdry.size()/2); + } + + public: + void Include (T xmin_ref, T xmax_ref) { + + // Determine the indices of xmin_ref & xmax_ref + int xmin_reg = -1; + int xmax_reg = -1; + for (int i = 0; i < bdry.size(); ++i) { + if ((i+1) % 2 && bdry[i] <= xmin_ref) xmin_reg++; + if (i % 2 && bdry[i] < xmin_ref) xmin_reg++; + if ((i+1) % 2 && bdry[i] <= xmax_ref) xmax_reg++; + if (i % 2 && bdry[i] < xmax_ref) xmax_reg++; + } + + //cout << "Include: xmin_reg = " << xmin_reg << "\txmax_reg = " << xmax_reg << endl; + + Vect new_bdry(bdry.size() + 2 * ((xmin_reg % 2 && xmax_reg % 2) - (xmax_reg - xmin_reg)/2)); + + int ishift = 0; + for (int i = 0; i <= xmin_reg; ++i) new_bdry[i] = bdry[i]; + if (xmin_reg % 2) { + new_bdry[xmin_reg + 1] = xmin_ref; + ishift++; + if (xmax_reg % 2) { + new_bdry[xmin_reg + 2] = xmax_ref; + ishift++; + } + } + else if ((xmin_reg + 1) % 2 && xmax_reg % 2) { + new_bdry[xmin_reg + 1] = xmax_ref; + ishift++; + } + for (int i = xmin_reg + ishift + 1; i < new_bdry.size(); ++i) + new_bdry[i] = bdry[xmax_reg - xmin_reg - ishift + i]; + + bdry = new_bdry; + + return; + } + + public: + void Exclude (T xmin_ref, T xmax_ref) { + + // Determine the indices of xmin_ref & xmax_ref + int xmin_reg = -1; + int xmax_reg = -1; + for (int i = 0; i < bdry.size(); ++i) { + if ((i+1) % 2 && bdry[i] <= xmin_ref) xmin_reg++; + if (i % 2 && bdry[i] < xmin_ref) xmin_reg++; + if ((i+1) % 2 && bdry[i] <= xmax_ref) xmax_reg++; + if (i % 2 && bdry[i] < xmax_ref) xmax_reg++; + } + + //cout << "Exclude: xmin_reg = " << xmin_reg << "\txmax_reg = " << xmax_reg << endl; + + Vect new_bdry(bdry.size() + 2 * (((xmin_reg + 1) % 2 && (xmax_reg + 1) % 2) - (xmax_reg - xmin_reg)/2)); + + int ishift = 0; + for (int i = 0; i <= xmin_reg; ++i) new_bdry[i] = bdry[i]; + if ((xmin_reg + 1) % 2) { + new_bdry[xmin_reg + 1] = xmin_ref; + ishift++; + if ((xmax_reg + 1) % 2) { + new_bdry[xmin_reg + 2] = xmax_ref; + ishift++; + } + } + else if (xmin_reg % 2 && (xmax_reg + 1) % 2) { + new_bdry[xmin_reg + 1] = xmax_ref; + ishift++; + } + for (int i = xmin_reg + ishift + 1; i < new_bdry.size(); ++i) + new_bdry[i] = bdry[xmax_reg - xmin_reg - ishift + i]; + + bdry = new_bdry; + + return; + } + + }; + + template + std::ostream& operator<< (std::ostream& s, Domain dom) + { + for (int i = 0; i < dom.Ndomains(); ++i) { + if (i > 0) s << endl; + s << dom.xmin(i) << "\t" << dom.xmax(i); + } + return(s); + } + + // ********************************* struct I_table ************************ + + struct I_table { + + DP (*function) (DP, DP); + int Nvals; + DP rhomin; + DP rhomax; + DP alpha; + DP logalpha; + DP prec; + //Vect_DP rho_tbl; + //Vect_DP I_tbl; + DP* rho_tbl; + DP* I_tbl; + + I_table (DP (*function) (DP, DP), DP rhomin_ref, DP rhomax_ref, int Nvals_ref, DP req_prec); + DP Return_val (DP req_rho); + void Save (); + bool Load (DP rhomin_ref, DP rhomax_ref, int Nvals_ref, DP req_prec); + + }; + + + // ********************************* struct Integral_table ************************ + + struct Integral_table { + + DP (*function) (DP, DP, int); + int Nvals; + DP rhomin; + DP rhomax; + DP alpha; + DP logalpha; + DP prec; + int maxnrpts; + DP* rho_tbl; + DP* I_tbl; + + Integral_table (DP (*function) (DP, DP, int), const char* filenameprefix_ref, DP rhomin_ref, DP rhomax_ref, int Nvals_ref, DP req_prec, int max_nr_pts); + DP Return_val (DP req_rho); + void Save (const char* filenameprefix); + bool Load (const char* filenameprefix, DP rhomin_ref, DP rhomax_ref, int Nvals_ref, DP req_prec, int max_nr_pts); + + }; + + + // ******************************** Recursive integration functions ****************************** + + DP Integrate_Riemann (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin, DP xmax, int Npts); + DP Integrate_Riemann_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, int Npts); + + DP Integrate_rec (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin, DP xmax, DP req_prec, int max_rec_level); + DP Integrate_rec_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int max_rec_level); + DP Integrate_rec_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int max_rec_level, ofstream& outfile); + DP Integrate_rec_using_table_and_file (DP (*function) (Vect_DP, I_table, ofstream&), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int max_rec_level, ofstream& outfile); + + + + // ******************************** Recursive version: optimal ****************************** + + struct data_pt { + DP x; + DP f; + DP dx; + }; + + struct Integral_result { + DP integ_est; + DP abs_prec; + DP rel_prec; + int n_vals; + }; + + std::ostream& operator<< (std::ostream& s, const Integral_result& res); + + class Integral_data { + + private: + data_pt* data; + DP* abs_d2f_dx; // second derivative * dx + DP max_abs_d2f_dx; // + + public: + //int n_vals; + Integral_result integ_res; + + public: + DP xmin; + DP xmax; + + public: + Integral_data (DP (*function_ref) (Vect_DP), Vect_DP& args, int arg_to_integ_ref, DP xmin_ref, DP xmax_ref); + Integral_data (DP (*function_ref) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ_ref, I_table Itable, DP xmin_ref, DP xmax_ref); + Integral_data (DP (*function_ref) (Vect_DP, Integral_table), Vect_DP& args, int arg_to_integ_ref, Integral_table Itable, DP xmin_ref, DP xmax_ref); + void Save (ofstream& outfile); + void Improve_estimate (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, int Npts_max); + void Improve_estimate (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, int Npts_max); + void Improve_estimate (DP (*function) (Vect_DP, Integral_table), Vect_DP& args, int arg_to_integ, Integral_table Itable, int Npts_max); + ~Integral_data (); + + }; + + Integral_result Integrate_optimal (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts); + Integral_result Integrate_optimal_using_table (DP (*function) (Vect_DP, I_table Itable), Vect_DP& args, int arg_to_integ, + I_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts); + Integral_result Integrate_optimal_using_table (DP (*function) (Vect_DP, Integral_table Itable), Vect_DP& args, int arg_to_integ, + Integral_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts); + + Integral_result Integrate_optimal_using_table (DP (*function) (Vect_DP, I_table Itable), Vect_DP& args, int arg_to_integ, + I_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts, ofstream& outfile); + + // ******************************** Recursive version: optimal, complex implementation ****************************** + + // NB: function returns complex values but takes real arguments + + struct data_pt_CX { + DP x; + complex f; + DP dx; + }; + + struct Integral_result_CX { + complex integ_est; + DP abs_prec; + DP rel_prec; + int n_vals; + }; + + class Integral_data_CX { + + private: + data_pt_CX* data; + DP* abs_d2f_dx; // second derivative * dx + DP max_abs_d2f_dx; // + + public: + //int n_vals; + Integral_result_CX integ_res; + + public: + DP xmin; + DP xmax; + + public: + Integral_data_CX (complex (*function_ref) (Vect_DP), Vect_DP& args, int arg_to_integ_ref, DP xmin_ref, DP xmax_ref); + //Integral_data_CX (complex (*function_ref) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ_ref, I_table Itable, DP xmin_ref, DP xmax_ref); + void Save (ofstream& outfile); + void Improve_estimate (complex (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, int Npts_max); + //void Improve_estimate (complex (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, int Npts_max); + ~Integral_data_CX (); + + }; + + Integral_result_CX Integrate_optimal (complex (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin, DP xmax, + DP req_rel_prec, DP req_abs_prec, int max_nr_pts); + //Integral_result_CX Integrate_optimal_using_table (DP (*function) (Vect_DP, I_table Itable), Vect_DP& args, int arg_to_integ, + // I_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts); + //Integral_result Integrate_optimal_using_table (DP (*function) (Vect_DP, I_table Itable), Vect_DP& args, int arg_to_integ, + // I_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts, ofstream& outfile); + + + + + +} // namespace JSC + +#endif diff --git a/include/JSC_LiebLin.h b/include/JSC_LiebLin.h new file mode 100644 index 0000000..e138afa --- /dev/null +++ b/include/JSC_LiebLin.h @@ -0,0 +1,190 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_LiebLin.h + +Purpose: Declares LiebLin gas-related classes and functions. + +***********************************************************/ + +#ifndef _LIEBLIN_ +#define _LIEBLIN_ + +#include "JSC.h" + +namespace JSC { + + // First, some global constants... + + //const DP ITER_REQ_PREC_LIEBLIN = 1.0e+6 * MACHINE_EPS_SQ; + const DP ITER_REQ_PREC_LIEBLIN = 1.0e+4 * MACHINE_EPS_SQ; + + const int LIEBLIN_Ix2_MIN = -1000000; // Like a UV cutoff. Assumption: never reached in scanning. + const int LIEBLIN_Ix2_MAX = -LIEBLIN_Ix2_MIN; + + //*********************************************************************** + + //class LiebLin_Bethe_State : public Bethe_State { + class LiebLin_Bethe_State { + + public: + DP c_int; // interaction parameter + DP L; + DP cxL; + int N; + string label; + //Vect OriginStateIx2; // quantum numbers of state on which excitations are built; always ordered + Vect Ix2_available; // quantum numbers which are allowable but not occupied + Vect index_first_hole_to_right; + Vect displacement; + Vect Ix2; + Vect lambdaoc; + //Vect BE; + Vect S; // scattering sum + Vect dSdlambdaoc; // its derivative + DP diffsq; + DP prec; + int conv; + int iter_Newton; + DP E; + int iK; + DP K; + DP lnnorm; + + public: + LiebLin_Bethe_State (); + LiebLin_Bethe_State (DP c_int_ref, DP L_ref, int N_ref); + LiebLin_Bethe_State& operator= (const LiebLin_Bethe_State& RefState); + + public: + int Charge () { return(N); }; + void Set_to_Label (string label_ref, const Vect& OriginStateIx2); + void Set_to_Label (string label_ref); // assumes OriginState == GroundState + void Set_Label_from_Ix2 (const Vect& OriginStateIx2); + void Set_Label_Internals_from_Ix2 (const Vect& OriginStateIx2); + bool Check_Admissibility(char whichDSF); // always returns true + void Find_Rapidities (bool reset_rapidities); + bool Check_Rapidities(); // checks that all rapidities are not nan + DP String_delta(); // trivially returns 0; exists to mirror spin chain function + bool Check_Symmetry (); // checks whether set of quantum numbers obeys { I } = { -I } + void Compute_lnnorm (); + void Compute_All (bool reset_rapidities); // solves BAE, computes E, K and lnnorm + void Set_Free_lambdaocs(); + void Iterate_BAE(DP damping); + void Iterate_BAE_S(DP damping); + void Iterate_BAE_Newton(DP damping); + void Compute_Energy (); + void Compute_Momentum (); + DP Kernel (int a, int b); + DP Kernel (DP lambdaoc_ref); + void Build_Reduced_Gaudin_Matrix (SQMat& Gaudin_Red); + void Build_Reduced_BEC_Quench_Gaudin_Matrix (SQMat& Gaudin_Red); + void Annihilate_ph_pair (int ipart, int ihole, const Vect& OriginStateIx2); + void Parity_Flip (); // takes all lambdaoc to -lambdaoc + inline bool Set_to_Inner_Skeleton (int iKneeded, const Vect& OriginIx2) + { + if (N < 3) JSCerror("N<3 incompatible with fixed momentum scanning"); + Ix2[0] = Ix2[1] - 2; + Ix2[N-1] = Ix2[N-2] + 2; + (*this).Compute_Momentum(); + if (iKneeded >= iK) Ix2[N-1] += 2*(iKneeded - iK); + else Ix2[0] += 2*(iKneeded - iK); + if (Ix2[0] < LIEBLIN_Ix2_MIN || Ix2[N-1] > LIEBLIN_Ix2_MAX) return(false); + (*this).Set_Label_from_Ix2 (OriginIx2); + return(true); + } + void Set_to_Outer_Skeleton (const Vect& OriginIx2) + { + Ix2[0] = LIEBLIN_Ix2_MIN + (N % 2) + 1; + Ix2[N-1] = LIEBLIN_Ix2_MAX - (N % 2) - 1; + (*this).Set_Label_from_Ix2 (OriginIx2); + //cout << "Set state to outer skeleton: Ix2 " << (*this).Ix2 << endl; + //cout << "label " << (*this).label << endl; + }; + }; + + inline bool Is_Inner_Skeleton (LiebLin_Bethe_State& State) { + return (State.N >= 2 && (State.Ix2[0] == State.Ix2[1] - 2 || State.Ix2[State.N-1] == State.Ix2[State.N-2] + 2)); + }; + inline bool Is_Outer_Skeleton (LiebLin_Bethe_State& State) { + return (State.N >= 2 && State.Ix2[0] == LIEBLIN_Ix2_MIN + (State.N % 2) + 1 && State.Ix2[State.N-1] == LIEBLIN_Ix2_MAX - (State.N % 2) - 1); + }; + inline bool Is_Outer_Skeleton (const LiebLin_Bethe_State& State) { + return (State.N >= 2 && State.Ix2[0] == LIEBLIN_Ix2_MIN + (State.N % 2) + 1 && State.Ix2[State.N-1] == LIEBLIN_Ix2_MAX - (State.N % 2) - 1); + }; + + inline bool Force_Descent (char whichDSF, LiebLin_Bethe_State& ScanState, LiebLin_Bethe_State& RefState, int desc_type_required, int iKmod, DP Chem_Pot) + { + bool forcedesc = false; + + // Force descent if we're computing density-density, we're at zero momentum and we're descending with momentum preserved: + if (whichDSF == 'd' && ScanState.iK == RefState.iK && desc_type_required > 8) forcedesc = true; + + // For BEC to c > 0 quench, g2(x=0): force first step + else if (whichDSF == 'B' && ScanState.label == RefState.label) forcedesc = true; + else if (whichDSF == 'C' && ScanState.label == RefState.label) forcedesc = true; + + return(forcedesc); + } + + std::ostream& operator<< (std::ostream& s, const LiebLin_Bethe_State& state); + + // FUNCTION DECLARATIONS: + + DP Chemical_Potential (LiebLin_Bethe_State& RefState); + DP Sumrule_Factor (char whichDSF, LiebLin_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax); + void Evaluate_F_Sumrule (char whichDSF, const LiebLin_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax, const char* RAW_Cstr, const char* FSR_Cstr); + void Evaluate_F_Sumrule (string prefix, char whichDSF, const LiebLin_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax); + void Evaluate_F_Sumrule (char whichDSF, DP c_int, DP L, int N, DP kBT, int nstates_req, DP Chem_Pot, int iKmin, int iKmax, const char* FSR_Cstr); + + // in LiebLin_Utils.cc + DP LiebLin_dE0_dc (DP c_int, DP L, int N); + DP LiebLin_vs (DP c_int, DP L, int N); + DP LiebLin_Dressed_Charge_N (DP c_int, DP L, int N); + int Momentum_Right_Excitations (LiebLin_Bethe_State& ScanState); + int Momentum_Left_Excitations (LiebLin_Bethe_State& ScanState); + DP ln_Overlap_with_BEC (LiebLin_Bethe_State& lambda); + + DP Particle_Hole_Excitation_Cost (char whichDSF, LiebLin_Bethe_State& AveragingState); + + complex ln_Density_ME (LiebLin_Bethe_State& lstate, LiebLin_Bethe_State& rstate); + complex ln_Psi_ME (LiebLin_Bethe_State& lstate, LiebLin_Bethe_State& rstate); + complex ln_g2_ME (LiebLin_Bethe_State& mu, LiebLin_Bethe_State& lambda); + + //DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, LiebLin_Bethe_State& LeftState, + // LiebLin_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile); + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, LiebLin_Bethe_State& LeftState, + LiebLin_Bethe_State& RefState, DP Chem_Pot, stringstream& DAT_outfile); + + DP LiebLin_Twisted_lnnorm (Vect >& lambdaoc, double cxL); + complex LiebLin_Twisted_ln_Overlap (DP expbeta, Vect lstate_lambdaoc, DP lstate_lnnorm, LiebLin_Bethe_State& rstate); + complex LiebLin_Twisted_ln_Overlap (complex expbeta, Vect > lstate_lambdaoc, DP lstate_lnnorm, LiebLin_Bethe_State& rstate); + complex LiebLin_ln_Overlap (Vect lstate_lambdaoc, DP lstate_lnnorm, LiebLin_Bethe_State& rstate); + complex LiebLin_ln_Overlap (Vect > lstate_lambdaoc, DP lstate_lnnorm, LiebLin_Bethe_State& rstate); + + // In src/LiebLin_Tgt0.cc: + //DP Entropy_rho (LiebLin_Bethe_State& RefState, int Delta); + //DP Entropy_Fixed_Delta (LiebLin_Bethe_State& RefState, int Delta); + //DP Entropy (LiebLin_Bethe_State& RefState, int Delta); + DP Entropy (LiebLin_Bethe_State& RefState); + //DP Canonical_Free_Energy (LiebLin_Bethe_State& RefState, DP kBT, int Delta); + DP Canonical_Free_Energy (LiebLin_Bethe_State& RefState, DP kBT); + //DP Entropy (LiebLin_Bethe_State& RefState, DP epsilon); + //DP Canonical_Free_Energy (LiebLin_Bethe_State& RefState, DP kBT, DP epsilon); + //LiebLin_Bethe_State Canonical_Saddle_Point_State (DP c_int, DP L, int N, DP kBT, int Delta); + //LiebLin_Bethe_State Canonical_Saddle_Point_State (DP c_int, DP L, int N, DP kBT, DP epsilon); + LiebLin_Bethe_State Canonical_Saddle_Point_State (DP c_int, DP L, int N, DP kBT); + LiebLin_Bethe_State Add_Particle_at_Center (const LiebLin_Bethe_State& RefState); + LiebLin_Bethe_State Remove_Particle_at_Center (const LiebLin_Bethe_State& RefState); + DP rho_of_lambdaoc_1 (LiebLin_Bethe_State& RefState, DP lambdaoc, DP delta); + DP rho_of_lambdaoc_2 (LiebLin_Bethe_State& RefState, DP lambdaoc, DP delta); + + +} // namespace JSC + +#endif diff --git a/include/JSC_Matrix.h b/include/JSC_Matrix.h new file mode 100644 index 0000000..3108637 --- /dev/null +++ b/include/JSC_Matrix.h @@ -0,0 +1,440 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_Matrix.h + +Purpose: Declares square matrix class. + +***********************************************************/ + +#ifndef _MATRIX_ +#define _MATRIX_ + +namespace JSC { + + + // CLASS DEFINITIONS + + template + class SQMat { + + private: + int dim; + T** M; + + public: + SQMat (int N); // initializes all elements of this n by n matrix to zero + SQMat (const SQMat& rhs); // copy constructor + SQMat (const T& a, int N); // initialize to diagonal matrix with value a (NOT like in NR !!!) + SQMat (const SQMat& a, const SQMat& b); // initialize to tensor product of a and b + SQMat (const SQMat& a, int row_id, int col_id); // init by cutting row row_id and col col_id + void Print (); + SQMat& operator= (const SQMat& rhs); // assignment + SQMat& operator= (const T& a); // assign 1 to diagonal elements (NOT like in NR !!!) + inline T* operator[] (const int i); // subscripting: pointer to row i + inline const T* operator[] (const int i) const; + SQMat& operator+= (const T& a); + SQMat& operator+= (const SQMat& a); + SQMat& operator-= (const T& a); + SQMat& operator-= (const SQMat& a); + SQMat& operator*= (const T& a); + SQMat& operator*= (const SQMat& a); + inline int size() const; + ~SQMat(); + + }; + + template + SQMat::SQMat (int N) : dim(N) , M(new T*[N]) + { + M[0] = new T[N*N]; + for (int i = 1; i < N; i++) M[i] = M[i-1] + N; + } + + template + SQMat::SQMat (const SQMat& rhs) : dim(rhs.dim) , M(new T*[dim]) + { + int i,j; + M[0] = new T[dim*dim]; + for (i = 1; i < dim; i++) M[i] = M[i-1] + dim; + for (i = 0; i < dim; i++) + for (j = 0; j < dim; j++) M[i][j] = rhs[i][j]; + } + + template + SQMat::SQMat (const T& a, int N) : dim(N) , M(new T*[dim]) + { + int i, j; + M[0] = new T[dim*dim]; + for (i = 1; i < dim; i++) M[i] = M[i-1] + dim; + for (i = 0; i < dim; i++) { + for (j = 0; j < dim; j++) M[i][j] = T(0); + M[i][i] = a; + } + } + + template + SQMat::SQMat (const SQMat& a, const SQMat& b) : dim (a.dim * b.dim) , M(new T*[a.dim * b.dim]) + { + M[0] = new T[a.dim * b.dim * a.dim * b.dim]; + + for (int i = 1; i < a.dim * b.dim; ++i) M[i] = M[i-1] + a.dim * b.dim; + + for (int i1 = 0; i1 < a.dim; ++i1) { + + for (int i2 = 0; i2 < a.dim; ++i2) { + + for (int j1 = 0; j1 < b.dim; ++j1) { + + for (int j2 = 0; j2 < b.dim; ++j2) { + + M[i1 * (b.dim) + j1][i2 * (b.dim) + j2] = a[i1][i2] * b[j1][j2]; + } + } + } + } + } + + template + SQMat::SQMat (const SQMat&a, int row_id, int col_id) : dim (a.dim - 1) , M(new T*[dim]) + { + if (dim == 0) { + JSCerror("Error: chopping a row and col from size one matrix."); + exit(1); + } + + M[0] = new T[dim * dim]; + + for (int i = 1; i < dim; ++i) M[i] = M[i-1] + dim; + + for (int i = 0; i < row_id; ++i) + for (int j = 0; j < col_id; ++j) M[i][j] = a[i][j]; + for (int i = row_id; i < dim; ++i) + for (int j = 0; j < col_id; ++j) M[i][j] = a[i+1][j]; + for (int i = 0; i < row_id; ++i) + for (int j = col_id; j < dim; ++j) M[i][j] = a[i][j+1]; + for (int i = row_id; i < dim; ++i) + for (int j = col_id; j < dim; ++j) M[i][j] = a[i+1][j+1]; + + } + + // operators + template + void SQMat::Print () + { + cout << endl; + for (int i = 0; i < dim; ++i) { + for (int j = 0; j < dim; ++j) cout << M[i][j] << " "; + cout << endl; + } + cout << endl; + } + + template + SQMat& SQMat::operator= (const SQMat& rhs) + { + if (this != &rhs) { + if (dim != rhs.dim) { + JSCerror("Assignment between matrices of different dimensions. Bailing out."); + exit(1); + } + + for (int i = 0; i < dim; ++i) + for (int j = 0; j < dim; ++j) M[i][j] = rhs[i][j]; + } + return *this; + } + + template + SQMat& SQMat::operator= (const T& a) + { + for (int i = 0; i < dim; ++i) { + for (int j = 0; j < dim; ++j) + M[i][j] = T(0); + M[i][i] = a; + } + return *this; + } + + template + inline T* SQMat::operator[] (const int i) + { + return M[i]; + } + + template + inline const T* SQMat::operator[] (const int i) const + { + return M[i]; + } + + template + SQMat& SQMat::operator+= (const T& a) + { + + for (int i = 0; i < dim; ++i) M[i][i] += a; + + return *this; + } + + template + SQMat& SQMat::operator+= (const SQMat& a) + { + if (dim != a.dim) { + JSCerror("Incompatible matrix sizes in matrix operator +."); + exit(1); + } + for (int i = 0; i < dim; ++i) { + for (int j = 0; j < dim; ++j) { + M[i][j] += a[i][j]; + } + } + return *this; + } + + template + SQMat& SQMat::operator-= (const T& a) + { + + for (int i = 0; i < dim; ++i) M[i][i] -= a; + + return *this; + } + + template + SQMat& SQMat::operator-= (const SQMat& a) + { + if (dim != a.dim) { + JSCerror("Incompatible matrix sizes in matrix operator +."); + exit(1); + } + for (int i = 0; i < dim; ++i) { + for (int j = 0; j < dim; ++j) { + M[i][j] -= a[i][j]; + } + } + return *this; + } + + template + SQMat& SQMat::operator*= (const T& a) + { + + for (int i = 0; i < dim; ++i) for (int j = 0; j < dim; ++j) M[i][j] *= a; + + return *this; + } + + template + SQMat& SQMat::operator*= (const SQMat& a) + { + + if (dim != a.dim) { + JSCerror("Incompatible matrix sizes in matrix operator *."); + exit(1); + } + + SQMat leftarg(*this); // use copy constructor. + + for (int i = 0; i < dim; ++i) { + + for (int j = 0; j < dim; ++j) { + + M[i][j] = 0.0; + + for (int k = 0; k < dim; ++k) { + + M[i][j] += leftarg[i][k] * a[k][j]; + } + } + } + + return *this; + } + + template + inline int SQMat::size() const + { + return dim; + } + + template + SQMat::~SQMat() + { + if (M != 0) { + delete[] (M[0]); + delete[] (M); + } + } + + + //***************************** + + template + class RecMat { + + private: + int nrows; + int ncols; + T** M; + + public: + RecMat (int Nrows, int Ncols); // initializes all elements of this n by n matrix to zero + RecMat (const T& a, int Nrows, int Ncols); + RecMat (const RecMat& rhs); // copy constructor + void Print (); + RecMat& operator= (const RecMat& rhs); // assignment + inline T* operator[] (const int i); // subscripting: pointer to row i + inline const T* operator[] (const int i) const; + inline int nr_rows() const; + inline int nr_cols() const; + ~RecMat(); + + }; + + template + RecMat::RecMat (int Nrows, int Ncols) : nrows(Nrows), ncols(Ncols), M(new T*[Nrows]) + { + M[0] = new T[Nrows*Ncols]; + for (int i = 1; i < Nrows; i++) M[i] = M[i-1] + Ncols; + + for (int i = 0; i < Nrows; i++) for (int j = 0; j < Ncols; j++) M[i][j] = T(0); + } + + template + RecMat::RecMat (const T& a, int Nrows, int Ncols) : nrows(Nrows), ncols(Ncols), M(new T*[Nrows]) + { + M[0] = new T[Nrows*Ncols]; + for (int i = 1; i < Nrows; i++) M[i] = M[i-1] + Ncols; + + for (int i = 0; i < Nrows; i++) for (int j = 0; j < Ncols; j++) { + if (i == j) M[i][i] = a; + else M[i][j] = T(0); + } + } + + template + RecMat::RecMat (const RecMat& rhs) : nrows(rhs.nrows), ncols(rhs.ncols), M(new T*[nrows]) + { + int i,j; + M[0] = new T[nrows*ncols]; + for (i = 1; i < nrows; i++) M[i] = M[i-1] + ncols; + for (i = 0; i < nrows; i++) + for (j = 0; j < ncols; j++) M[i][j] = rhs[i][j]; + } + + // operators + template + void RecMat::Print () + { + cout << endl; + for (int i = 0; i < nrows; ++i) { + for (int j = 0; j < ncols; ++j) cout << M[i][j] << " "; + cout << endl; + } + cout << endl; + } + + template + RecMat& RecMat::operator= (const RecMat& rhs) + { + if (this != &rhs) { + if (nrows != rhs.nrows || ncols != rhs.ncols) { + if (M != 0) { + delete[] (M[0]); + delete[] (M); + } + nrows = rhs.nrows; + ncols = rhs.ncols; + M = new T*[nrows]; + M[0] = new T[nrows * ncols]; + } + + for (int i = 0; i < nrows; ++i) + for (int j = 0; j < ncols; ++j) M[i][j] = rhs[i][j]; + } + return *this; + } + + template + inline T* RecMat::operator[] (const int i) + { + return M[i]; + } + + template + inline const T* RecMat::operator[] (const int i) const + { + return M[i]; + } + + template + inline int RecMat::nr_rows() const + { + return nrows; + } + + template + inline int RecMat::nr_cols() const + { + return ncols; + } + + template + inline std::ostream& operator<< (std::ostream& s, const RecMat& matrix) + { + for (int i = 0; i < matrix.nr_rows(); ++i) { + for (int j = 0; j < matrix.nr_cols(); ++j) s << matrix[i][j] << " "; + s << endl; + } + + return (s); + } + + template + RecMat::~RecMat() + { + if (M != 0) { + delete[] (M[0]); + delete[] (M); + } + } + + // TYPEDEFS: + + typedef JSC::SQMat SQMat_DP; + typedef JSC::SQMat > SQMat_CX; + + + // FUNCTION DEFINITIONS + + // Functions in src/MATRIX directory + + DP det_LU (SQMat_DP a); + DP lndet_LU (SQMat_DP a); + complex lndet_LU_dstry (SQMat_DP& a); + complex det_LU_CX (SQMat_CX a); + complex lndet_LU_CX (SQMat_CX a); + complex lndet_LU_CX_dstry (SQMat_CX& a); + void eigsrt (Vect_DP& d, SQMat_DP& v); + void balanc (SQMat_DP& a); + void elmhes (SQMat_DP& a); + void gaussj (SQMat_DP& a, SQMat_DP& b); + void hqr (SQMat_DP& a, Vect_CX& wri); + void jacobi (SQMat_DP& a, Vect_DP& d, SQMat_DP& v, int& nrot); + void lubksb (SQMat_DP& a, Vect_INT& indx, Vect_DP& b); + void lubksb_CX (SQMat_CX& a, Vect_INT& indx, Vect_CX& b); + void ludcmp (SQMat_DP& a, Vect_INT& indx, DP& d); + void ludcmp_CX (SQMat_CX& a, Vect_INT& indx, DP& d); + DP pythag(DP a, DP b); + void tqli(Vect_DP& d, Vect_DP& e, SQMat_DP& z); + void tred2 (SQMat_DP& a, Vect_DP& d, Vect_DP& e); + +} // namespace JSC + +#endif diff --git a/include/JSC_NRG.h b/include/JSC_NRG.h new file mode 100644 index 0000000..31776ff --- /dev/null +++ b/include/JSC_NRG.h @@ -0,0 +1,36 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_NRG.h + +Purpose: Declares NRG-related classes and functions. + +***********************************************************/ + +#ifndef _NRG_ +#define _NRG_ + +#include "JSC.h" + +namespace JSC { + + DP K_Weight_integrand (Vect_DP args); // weighing function for state selection + + //void Select_States_for_NRG (DP c_int, DP L, int N, int iKmin, int iKmax, int Nstates_required, bool symmetric_states, int iKmod, + // int weighing_option, DP (*weight_integrand_fn) (Vect_DP), Vect_DP& args_to_weight_integrand); + void Select_States_for_NRG (DP c_int, DP L, int N, int iKmin, int iKmax, int Nstates_required, bool symmetric_states, int iKmod, + //int weighing_option, DP (*weight_integrand_fn) (Vect_DP), Vect_DP& args_to_weight_integrand) + int weighing_option, Vect >& FT_of_potential); + + void Build_DFF_Matrix_Block_for_NRG (DP c_int, DP L, int N, int iKmin, int iKmax, int Nstates_required, bool symmetric_states, int iKmod, + int weighing_option, int label_left_begin, int label_left_end, int label_right_begin, int label_right_end, + int block_option, DP* DFF_block_1, DP* DFF_block_2, Vect_DP Kweight); + +} + +#endif // _NRG_ diff --git a/include/JSC_ODSLF.h b/include/JSC_ODSLF.h new file mode 100644 index 0000000..76d6350 --- /dev/null +++ b/include/JSC_ODSLF.h @@ -0,0 +1,463 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Heis.h + +Purpose: Declares lattice spinless fermion classes and functions. + +***********************************************************/ + +#ifndef _ODSLF_ +#define _ODSLF_ + +#include "JSC.h" + +namespace JSC { + + //**************************************************************************** + + // Objects in class ODSLF_Base are a checked vector containing the number of rapidities of allowable types for a given state + + class ODSLF_Base { + + public: + int Mdown; // total number of down spins + Vect Nrap; // Nrap[i] contains the number of rapidities of type i, i = 0, Nstrings - 1. + int Nraptot; // total number of strings in this state + Vect Ix2_infty; // Ix2_infty[i] contains the max of BAE function for the (half-)integer I[i], i = 0, Nstrings - 1. + Vect Ix2_max; // Ix2_max[i] contains the integer part of 2*I_infty, with correct parity for base. + long long int id; // identification number + + public: + ODSLF_Base (); + ODSLF_Base (const ODSLF_Base& RefBase); // copy constructor + ODSLF_Base (const Heis_Chain& RefChain, int M); // constructs configuration with all Mdown in one-string of +1 parity + ODSLF_Base (const Heis_Chain& RefChain, const Vect& Nrapidities); // sets to Nrapidities vector, and checks consistency + ODSLF_Base (const Heis_Chain& RefChain, long long int id_ref); + inline int& operator[] (const int i); + inline const int& operator[] (const int i) const; + ODSLF_Base& operator= (const ODSLF_Base& RefBase); + bool operator== (const ODSLF_Base& RefBase); + bool operator!= (const ODSLF_Base& RefBase); + + void Compute_Ix2_limits(const Heis_Chain& RefChain); // computes the Ix2_infty and Ix2_max + + void Scan_for_Possible_Types (Vect& possible_type_id, int& nfound, int base_level, Vect& Nexcitations); + Vect Possible_Types (); // returns a vector of possible types + + }; + + inline int& ODSLF_Base::operator[] (const int i) + { + return Nrap[i]; + } + + inline const int& ODSLF_Base::operator[] (const int i) const + { + return Nrap[i]; + } + + //**************************************************************************** + + // Objects in class ODSLF_Ix2_Config carry all the I's of a given state + + class ODSLF_Ix2_Config { + + //private: + public: + int Nstrings; + Vect Nrap; + int Nraptot; + + int** Ix2; + + //Vect > Ix2; + + public: + ODSLF_Ix2_Config (); + ODSLF_Ix2_Config (const Heis_Chain& RefChain, int M); // constructor, puts I's to ground state + ODSLF_Ix2_Config (const Heis_Chain& RefChain, const ODSLF_Base& base); // constructor, putting I's to lowest-energy config + // consistent with Heis_Base configuration for chain RefChain + ODSLF_Ix2_Config& operator= (const ODSLF_Ix2_Config& RefConfig); + inline int* operator[] (const int i); + //inline Vect operator[] (const int i); + inline const int* operator[] (const int i) const; + //inline const Vect operator[] (const int i) const; + ~ODSLF_Ix2_Config(); + }; + + inline int* ODSLF_Ix2_Config::operator[] (const int i) + //inline Vect Ix2_Config::operator[] (const int i) + { + return Ix2[i]; + } + + inline const int* ODSLF_Ix2_Config::operator[] (const int i) const + //inline const Vect Ix2_Config::operator[] (const int i) const + { + return Ix2[i]; + } + + std::ostream& operator<< (std::ostream& s, const ODSLF_Ix2_Config& RefConfig); + + //**************************************************************************** + + // Objects in class ODSLF_Lambda carry all rapidities of a state + + class ODSLF_Lambda { + + private: + int Nstrings; + Vect Nrap; + int Nraptot; + DP** lambda; + //Vect > lambda; + + public: + ODSLF_Lambda (); + ODSLF_Lambda (const Heis_Chain& RefChain, int M); // constructor, puts all lambda's to zero + ODSLF_Lambda (const Heis_Chain& RefChain, const ODSLF_Base& base); // constructor, putting I's to lowest-energy config + // consistent with Heis_Base configuration for chain RefChain + ODSLF_Lambda& operator= (const ODSLF_Lambda& RefConfig); + inline DP* operator[] (const int i); + //inline Vect operator[] (const int i); + inline const DP* operator[] (const int i) const; + //inline const Vect operator[] (const int i) const; + ~ODSLF_Lambda(); + + }; + + inline DP* ODSLF_Lambda::operator[] (const int i) + //inline Vect Lambda::operator[] (const int i) + { + return lambda[i]; + } + + inline const DP* ODSLF_Lambda::operator[] (const int i) const + //inline const Vect Lambda::operator[] (const int i) const + { + return lambda[i]; + } + + + //**************************************************************************** + + // Objects in class ODSLF_Ix2_Offsets carry Young tableau representations of the Ix2 configurations + + class ODSLF_Ix2_Offsets { + + public: + ODSLF_Base base; + Vect Tableau; // vector of pointers to tableaux at each level + long long int type_id; + long long int id; // id number of offset + long long int maxid; // max id number allowable + + public: + ODSLF_Ix2_Offsets (); + ODSLF_Ix2_Offsets (const ODSLF_Ix2_Offsets& RefOffset); // copy constructor + ODSLF_Ix2_Offsets (const ODSLF_Base& RefBase, long long int req_type_id); + ODSLF_Ix2_Offsets (const ODSLF_Base& RefBase, Vect nparticles); // sets all tableaux to empty ones, with nparticles[] at each level + ODSLF_Ix2_Offsets& operator= (const ODSLF_Ix2_Offsets& RefOffset); + bool operator<= (const ODSLF_Ix2_Offsets& RefOffsets); + bool operator>= (const ODSLF_Ix2_Offsets& RefOffsets); + + public: + void Set_to_id (long long int idnr); + void Compute_id (); + void Compute_type_id (); + + public: + bool Add_Boxes_From_Lowest (int Nboxes, bool odd_sectors); // adds Nboxes in minimal energy config, all boxes in either even or odd sectors + + }; + + inline long long int ODSLF_Ix2_Offsets_type_id (Vect& nparticles) + { + long long int type_id_here = 0ULL; + + for (int i = 0; i < nparticles.size(); ++i) + type_id_here += nparticles[i] * pow_ulli(10ULL, i); + + return(type_id_here); + } + + //**************************************************************************** + // Objects in class ODSLF_Ix2_Offsets_List carry a vector of used Ix2_Offsets + + class ODSLF_Ix2_Offsets_List { + + public: + int ndef; + Vect Offsets; + + public: + ODSLF_Ix2_Offsets_List (); + ODSLF_Ix2_Offsets& Return_Offsets (ODSLF_Base& RefBase, Vect nparticles); // returns the Ix2_Offsets corresponding to nparticles[]/base + ODSLF_Ix2_Offsets& Return_Offsets (ODSLF_Base& RefBase, long long int req_type_id); + }; + + + //**************************************************************************** + + // Objects in class ODSLF_Bethe_State carry all information about an eigenstate + + // Derived classes include XXZ_Bethe_State, XXX_Bethe_State, XXZ_gpd_Bethe_State + // These contain subclass-specific functions and data. + + class ODSLF_Bethe_State { + + public: + Heis_Chain chain; + ODSLF_Base base; + ODSLF_Ix2_Offsets offsets; + ODSLF_Ix2_Config Ix2; + ODSLF_Lambda lambda; + ODSLF_Lambda BE; // Bethe equation for relevant rapidity, in the form BE = theta - (1/N)\sum ... - \pi I/N = 0 + DP diffsq; // sum of squares of rapidity differences in last iteration + int conv; // convergence status + int iter; // number of iterations necessary for convergence + int iter_Newton; // number of iterations necessary for convergence (Newton method) + DP E; // total energy + int iK; // K = 2.0*PI * iK/Nsites + DP K; // total momentum + DP lnnorm; // ln of norm of reduced Gaudin matrix + //long long int id; + //long long int maxid; + long long int base_id; + long long int type_id; + long long int id; + long long int maxid; + int nparticles; + + public: + ODSLF_Bethe_State (); + ODSLF_Bethe_State (const ODSLF_Bethe_State& RefState); // copy constructor + ODSLF_Bethe_State (const ODSLF_Bethe_State& RefState, long long int type_id_ref); // new state with requested type_id + ODSLF_Bethe_State (const Heis_Chain& RefChain, int M); // constructor to ground-state configuration + ODSLF_Bethe_State (const Heis_Chain& RefChain, const ODSLF_Base& base); // constructor to lowest-energy config with base + ODSLF_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref); + virtual ~ODSLF_Bethe_State () {}; + + public: + int Charge () { return(base.Mdown); }; + //void Set_I_Offset (const I_Offset& RefOffset); // sets the Ix2 to given offsets + void Set_Ix2_Offsets (const ODSLF_Ix2_Offsets& RefOffset); // sets the Ix2 to given offsets + void Set_to_id (long long int id_ref); + void Set_to_id (long long int id_ref, ODSLF_Bethe_State& RefState); + int Nparticles (); // counts the number of particles in state once Ix2 offsets set (so type_id is correctly set) + bool Check_Symmetry (); // checks whether the I's are symmetrically distributed + void Compute_diffsq (); // \sum BE[j][alpha]^2 + void Iterate_BAE (); // Finds new set of lambda[j][alpha] from previous one by simple iteration + void Iterate_BAE_Newton (); // Finds new set of lambda[j][alpha] from previous one by a Newton step + void Find_Rapidities (bool reset_rapidities); // Finds the rapidities + void Find_Rapidities_Twisted (bool reset_rapidities, DP twist); // Finds the rapidities with twist added to RHS of logBE + void BAE_smackdown (DP max_allowed); + void Solve_BAE_smackdown (DP max_allowed, int maxruns); + void Solve_BAE (int j, int alpha, DP req_prec, int itermax); + void Solve_BAE_interp (DP interp_prec, int max_iter_interp); + void Solve_BAE_straight_iter (DP interp_prec, int max_iter_interp); + void Solve_BAE_Newton (DP Newton_prec, int max_iter_Newton); + void Compute_lnnorm (); + void Compute_Momentum (); + void Compute_All (bool reset_rapidities); // solves BAE, computes E, K and lnnorm + bool Boost_Momentum (int iKboost); + + // Virtual functions, all defined in the derived classes + + public: + virtual void Set_Free_lambdas() { JSCerror("ODSLF_Bethe_State::..."); } // sets the rapidities to solutions of BAEs without scattering terms + virtual bool Check_Admissibility(char option) { JSCerror("ODSLF_Bethe_State::..."); return(false); } + // verifies that we don't have a symmetrical Ix2 config with a Ix2 == 0 for a string of even length >= 2. + virtual void Compute_BE (int j, int alpha) { JSCerror("ODSLF_Bethe_State::..."); } + virtual void Compute_BE () { JSCerror("ODSLF_Bethe_State::..."); } + virtual DP Iterate_BAE(int i, int alpha) { JSCerror("ODSLF_Bethe_State::..."); return(0.0);} + virtual bool Check_Rapidities() { JSCerror("ODSLF_Bethe_State::..."); return(false); } + virtual void Compute_Energy () { JSCerror("ODSLF_Bethe_State::..."); } + virtual void Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red) { JSCerror("ODSLF_Bethe_State::..."); } + }; + + inline bool Force_Descent (char whichDSF, ODSLF_Bethe_State& ScanState, ODSLF_Bethe_State& RefState, int desc_type_required, int iKmod, DP Chem_Pot) + { + JSCerror("Need to implement Force_Descent properly for ODSLF."); + + bool force_descent = false; + + // Force descent if energy of ScanState is lower than that of RefState + if (ScanState.E - RefState.E - (ScanState.base.Mdown - RefState.base.Mdown) < 0.0) return(true); + /* + // We force descent if + // 1) - there exists a higher string whose quantum number is still on 0 + // AND - there is at most a single particle-hole in the 0 base level + // AND - either the particle or the hole hasn't yet moved. + if (RefState.base_id/100000LL > 0) { // there is a higher string + int type0 = RefState.type_id % 10000; + if (type0 == 0 + || type0 == 101 && RefState.offsets.Tableau[0].id * RefState.offsets.Tableau[2].id == 0LL + || type0 == 110 && RefState.offsets.Tableau[1].id * RefState.offsets.Tableau[2].id == 0LL + || type0 == 1001 && RefState.offsets.Tableau[0].id * RefState.offsets.Tableau[3].id == 0LL + || type0 == 1010 && RefState.offsets.Tableau[1].id * RefState.offsets.Tableau[3].id == 0LL) // single p-h pair in base level 0 + for (int j = 1; j < RefState.chain.Nstrings; ++j) { + if (RefState.base[j] == 1 && RefState.Ix2[j][0] == 0) { + force_descent = true; + } + } + } + */ + // Force descent if quantum nr distribution is symmetric: + if (RefState.Check_Symmetry()) force_descent = true; + + return(force_descent); + } + + std::ostream& operator<< (std::ostream& s, const ODSLF_Bethe_State& state); + + //**************************************************************************** + + // Objects in class XXZ_Bethe_State carry all extra information pertaining to XXZ gapless + + class ODSLF_XXZ_Bethe_State : public ODSLF_Bethe_State { + + public: + ODSLF_Lambda sinhlambda; + ODSLF_Lambda coshlambda; + ODSLF_Lambda tanhlambda; + + public: + ODSLF_XXZ_Bethe_State (); + ODSLF_XXZ_Bethe_State (const ODSLF_XXZ_Bethe_State& RefState); // copy constructor + ODSLF_XXZ_Bethe_State (const Heis_Chain& RefChain, int M); // constructor to ground-state configuration + ODSLF_XXZ_Bethe_State (const Heis_Chain& RefChain, const ODSLF_Base& base); // constructor to lowest-energy config with base + ODSLF_XXZ_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref); // constructor to lowest-energy config with bas + + public: + ODSLF_XXZ_Bethe_State& operator= (const ODSLF_XXZ_Bethe_State& RefState); + + public: + void Set_Free_lambdas(); // sets the rapidities to solutions of BAEs without scattering terms + void Compute_sinhlambda(); + void Compute_coshlambda(); + void Compute_tanhlambda(); + bool Check_Admissibility(char option); // verifies that we don't have a symmetrical Ix2 config with a Ix2 == 0 for a string of even length >= 2. + void Compute_BE (int j, int alpha); + void Compute_BE (); + DP Iterate_BAE(int i, int j); + bool Check_Rapidities(); // checks that all rapidities are not nan + void Compute_Energy (); + //void Compute_Momentum (); + void Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red); + + // XXZ specific functions: + public: + + }; + + //**************************************************************************** + /* + // Objects in class ODSLF_XXX_Bethe_State carry all extra information pertaining to XXX antiferromagnet + + class ODSLF_XXX_Bethe_State : public ODSLF_Bethe_State { + + public: + ODSLF_XXX_Bethe_State (); + ODSLF_XXX_Bethe_State (const ODSLF_XXX_Bethe_State& RefState); // copy constructor + ODSLF_XXX_Bethe_State (const Heis_Chain& RefChain, int M); // constructor to ground-state configuration + ODSLF_XXX_Bethe_State (const Heis_Chain& RefChain, const ODSLF__Base& base); // constructor to lowest-energy config with base + ODSLF_XXX_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref); // constructor to lowest-energy config with base + + public: + ODSLF_XXX_Bethe_State& operator= (const ODSLF_XXX_Bethe_State& RefState); + + public: + void Set_Free_lambdas(); // sets the rapidities to solutions of BAEs without scattering terms + bool Check_Admissibility(char option); // verifies that we don't have a symmetrical Ix2 config with a Ix2 == 0 for a string of even length >= 2. + void Compute_BE (int j, int alpha); + void Compute_BE (); + DP Iterate_BAE(int i, int j); + bool Check_Rapidities(); // checks that all rapidities are not nan + void Compute_Energy (); + //void Compute_Momentum (); + void Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red); + + // XXX specific functions + public: + bool Check_Finite_rap (); + + }; + */ + //**************************************************************************** + /* + // Objects in class ODSLF_XXZ_gpd_Bethe_State carry all extra information pertaining to XXZ gapped antiferromagnets + + class ODSLF_XXZ_gpd_Bethe_State : public ODSLF__Bethe_State { + + public: + Lambda sinlambda; + Lambda coslambda; + Lambda tanlambda; + + public: + ODSLF_XXZ_gpd_Bethe_State (); + ODSLF_XXZ_gpd_Bethe_State (const ODSLF_XXZ_gpd_Bethe_State& RefState); // copy constructor + ODSLF_XXZ_gpd_Bethe_State (const Heis_Chain& RefChain, int M); // constructor to ground-state configuration + ODSLF_XXZ_gpd_Bethe_State (const Heis_Chain& RefChain, const ODSLF_Base& base); // constructor to lowest-energy config with base + ODSLF_XXZ_gpd_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref); // constructor to lowest-energy config with base + + public: + ODSLF_XXZ_gpd_Bethe_State& operator= (const ODSLF_XXZ_gpd_Bethe_State& RefState); + + public: + void Set_Free_lambdas(); // sets the rapidities to solutions of BAEs without scattering terms + void Compute_sinlambda(); + void Compute_coslambda(); + void Compute_tanlambda(); + int Weight(); // weight function for contributions cutoff + bool Check_Admissibility(char option); // verifies that we don't have a symmetrical Ix2 config with a Ix2 == 0 for a string of even length >= 2. + void Compute_BE (int j, int alpha); + void Compute_BE (); + DP Iterate_BAE(int i, int j); + void Iterate_BAE_Newton(); + bool Check_Rapidities(); // checks that all rapidities are not nan and are in interval ]-PI/2, PI/2] + void Compute_Energy (); + //void Compute_Momentum (); + void Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red); + + // XXZ_gpd specific functions + public: + + }; + */ + //*********************************************** + + // Function declarations + /* + // in M_vs_H.cc + DP Ezero (DP Delta, int N, int M); + DP H_vs_M (DP Delta, int N, int M); + DP HZmin (DP Delta, int N, int M, Vect_DP& Ezero_ref); + int M_vs_H (DP Delta, int N, DP HZ); + + DP X_avg (char xyorz, DP Delta, int N, int M); + */ + DP Chemical_Potential (const ODSLF_Bethe_State& RefState); + //DP Sumrule_Factor (char whichDSF, Heis_Bethe_State& RefState, DP Chem_Pot, bool fixed_iK, int iKneeded); + DP Sumrule_Factor (char whichDSF, ODSLF_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax); + void Evaluate_F_Sumrule (string prefix, char whichDSF, const ODSLF_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax); + + complex ln_Sz_ME (ODSLF_XXZ_Bethe_State& A, ODSLF_XXZ_Bethe_State& B); + complex ln_Smin_ME (ODSLF_XXZ_Bethe_State& A, ODSLF_XXZ_Bethe_State& B); + + //DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, ODSLF_XXZ_Bethe_State& LeftState, + // ODSLF_XXZ_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile); + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, ODSLF_XXZ_Bethe_State& LeftState, + ODSLF_XXZ_Bethe_State& RefState, DP Chem_Pot, stringstream& DAT_outfile); + + +} // namespace JSC + +#endif diff --git a/include/JSC_Scan.h b/include/JSC_Scan.h new file mode 100644 index 0000000..be98d79 --- /dev/null +++ b/include/JSC_Scan.h @@ -0,0 +1,1263 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_Scan.h + +Purpose: Declares all classes and functions used in the + ABACUS logic of scanning with threads. + + +***********************************************************/ + +#ifndef _SCAN_ +#define _SCAN_ + +#include "JSC.h" + +namespace JSC { + + const int MAX_STATE_LIST_SIZE = 10000; + + // Functions in src/UTILS/Data_File_Name.cc: + void Data_File_Name (stringstream& name, char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, DP L2, string defaultname); + void Data_File_Name (stringstream& name, char whichDSF, int iKmin, int iKmax, DP kBT, LiebLin_Bethe_State& State, LiebLin_Bethe_State& RefScanState, string defaultname); + void Data_File_Name (stringstream& name, char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, DP kBT, int N2, string defaultname); + void Data_File_Name (stringstream& name, char whichDSF, int iKmin, int iKmax, DP kBT, Heis_Bethe_State& State, Heis_Bethe_State& RefScanState, string defaultname); + void ODSLF_Data_File_Name (stringstream& name, char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, DP kBT, int N2, string defaultname); + void Data_File_Name (stringstream& name, char whichDSF, int iKmin, int iKmax, DP kBT, ODSLF_Bethe_State& State, ODSLF_Bethe_State& RefScanState, string defaultname); + + // Coding to convert ints to strings: for application in reduced labels + //const int JSCcodingsize = 64; // use a multiple of 2 to accelerate divisions in labeling. + //const char JSCcoding[] = {'a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v', 'w', 'x', 'y', 'z', 'A', 'B', 'C', 'D', 'E', 'F', 'G', 'H', 'I', 'J', 'K', 'L', 'M', 'N', 'O', 'P', 'Q', 'R', 'S', 'T', 'U', 'V', 'W', 'X', 'Y', 'Z', '0', '1', '2', '3', '4', '5', '6', '7', '8', '9', '!', '?'}; + + // From ABACUS++T_8 onwards: forbid special characters as |, :, !, ? and all capital letters in labels. + // This is due to the capitalization-preserving but capitalization-insensitive HFS+ filesystem on Mac OS X. + const int JSCcodingsize = 32; + const char JSCcoding[] = {'0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v'}; + const char LABELSEP = '_'; // was _ + const char TYPESEP = 'x'; // was | + const char EXCSEP = 'y'; // was : + const char INEXCSEP = 'z'; // was @ + + struct State_Label_Data { + + //int ntypes; // how many types of particles are present in the state + Vect type; // integer type labels of the types present + Vect M; // how many particles of each type + Vect nexc; // how many excitations as compared to the reference state used + Vect > Ix2old; // which Ix2 will be excited + Vect > Ix2exc; // which Ix2 the excitation has shifted to + + State_Label_Data (const Vect& type_ref, const Vect& M_ref, + const Vect& nexc_ref, const Vect >& Ix2old_ref, const Vect >& Ix2exc_ref) + { + type = type_ref; M = M_ref; nexc = nexc_ref; Ix2old = Ix2old_ref; Ix2exc = Ix2exc_ref; + } + + }; + + // DEPRECATED ++G_5 + //State_Label_Data Read_State_Ulabel (string ulabel); + //string Return_State_Ulabel (State_Label_Data data); + //string Return_State_Ulabel (const Vect >& ScanIx2, const Vect >& OriginIx2); + //string Return_State_Ulabel (const Vect& ScanIx2, const Vect& OriginIx2); // if there is only one type + + string Extract_Base_Label (string label); // works for labels and complabels + string Extract_nexc_Label (string label); + + // For compressed labels: conversions between integers and char/strings. + string Convert_POSINT_to_STR (int int_to_convert); + int Convert_CHAR_to_POSINT (char char_to_convert); + int Convert_STR_to_POSINT (string str_to_convert); + + State_Label_Data Read_Base_Label (string label); + State_Label_Data Read_State_Label (string label, const Vect >& OriginIx2); + State_Label_Data Read_State_Label (string label, const Vect& OriginIx2); // if there is only one type + string Return_State_Label (State_Label_Data data, const Vect >& OriginIx2); + string Return_State_Label (State_Label_Data data, const Vect& OriginIx2); // if there is only one type + string Return_State_Label (const Vect >& ScanIx2, const Vect >& OriginIx2); + string Return_State_Label (const Vect& ScanIx2, const Vect& OriginIx2); // if there is only one type + Vect > Return_Ix2_from_Label (string label_ref, const Vect >& OriginIx2); + Vect Return_Ix2_from_Label (string label_ref, const Vect& OriginIx2); // specialization to Lieb-Liniger + //bool Set_to_Label (string label_ref, Vect >& Ix2, const Vect >& OriginIx2); + + // Functions for descending states: in SCAN/Descendents.cc + //Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2, const Vect& Ix2_min, const Vect& Ix2_max); + //Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2, const Vect& Ix2_min, const Vect& Ix2_max); + //Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2, const Vect& Ix2_min, const Vect& Ix2_max); + // Specialization for Lieb-Liniger case: + // ++G_5 + //Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const Vect& OriginIx2, const Vect& BaseScanIx2); + //Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const Vect& OriginIx2, const Vect& BaseScanIx2); + //Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const Vect& OriginIx2, const Vect& BaseScanIx2); + //Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const LiebLin_Bethe_State& OriginState); + //Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const LiebLin_Bethe_State& OriginState); + //Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const LiebLin_Bethe_State& OriginState); + //Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const Heis_Bethe_State& OriginState); + //Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const Heis_Bethe_State& OriginState); + //Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const Heis_Bethe_State& OriginState); + // ++G_6.0 + //Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc); + //Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc); + //Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc_up, bool disperse_only_current_exc_down); + //Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc); + //Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc); + //Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc_up, bool disperse_only_current_exc_down); + // ++G_6.1 + Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc); + Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc); + Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc_up, bool preserve_nexc_up, bool disperse_only_current_exc_down, bool preserve_nexc_down); + Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc); + Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc); + Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc_up, bool preserve_nexc_up, bool disperse_only_current_exc_down, bool preserve_nexc_down); + + // For symmetric state scanning: + Vect Descendent_States_with_iK_Stepped_Up_rightIx2only + (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc); + Vect Descendent_States_with_iK_Stepped_Down_rightIx2only + (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc); + Vect Descendent_States_with_iK_Stepped_Up_rightIx2only + (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc); + Vect Descendent_States_with_iK_Stepped_Down_rightIx2only + (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc); + + + + // Conversion between labels and complabels: DEPRECATED ++G_5 + //string Compress_Ulabel (string ulabel_ref, const Vect >& OriginIx2); + //string Compress_Ulabel (string ulabel_ref, const Vect& OriginIx2); // if there is only one type + //string Uncompress_Label (string complabel_ref, const Vect >& OriginIx2); + //string Uncompress_Label (string complabel_ref, const Vect& OriginIx2); // if there is only one type + + // Functions in src/SCAN/General_Scan.cc: + //template + //void General_Scan (char whichDSF, Tstate& RefState, int Max_Secs, bool refine); + //void ScanLiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, + // int Max_Secs, DP target_sumrule, bool refine, int rank, int nr_processors); + //void Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, int Max_Secs, bool refine); + //void Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKneeded, int Max_Secs, bool refine); + //void Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iK_UL, int Max_Secs, bool refine); + // Finite temperature: + //void Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, int Max_Secs, DP target_sumrule, bool refine, int rank, int nr_processors); + void Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors); + void Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, + int Max_Secs, DP target_sumrule, bool refine); + // Over arbitrary excited state: + //void Scan_LiebLin (char whichDSF, LiebLin_Bethe_State AveragingState, string defaultname, int iKmin, int iKmax, int Max_Secs, DP target_sumrule, bool refine, int rank, int nr_processors); + void Scan_LiebLin (char whichDSF, LiebLin_Bethe_State AveragingState, string defaultScanStatename, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors); + void Scan_LiebLin (char whichDSF, LiebLin_Bethe_State AveragingState, string defaultname, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine); + + //void Scan_Heis (char whichDSF, DP Delta, DP N, int M, bool fixed_iK, int iKneeded, int Max_Secs, bool refine, int rank, int nr_processors); + void Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors); + void Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine); + //void Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, int Max_Secs, bool refine); + //void Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKneeded, int Max_Secs, bool refine); + //void Scan_Heis (char whichDSF, DP Delta, int N, int M, int Max_Secs, bool refine); + + void Scan_Heis (char whichDSF, XXZ_Bethe_State& AveragingState, string defaultScanStatename, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors); + void Scan_Heis (char whichDSF, XXX_Bethe_State& AveragingState, string defaultScanStatename, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors); + void Scan_Heis (char whichDSF, XXZ_gpd_Bethe_State& AveragingState, string defaultScanStatename, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors); + + //void Scan_LiebLin_Geometric_Quench (DP c_int, DP L_1, int type_id_1, long long int id_1, DP L_2, int N, int iK_UL, int Max_Secs, bool refine); + void Scan_LiebLin_Geometric_Quench (DP c_int, DP L_1, int type_id_1, long long int id_1, DP L_2, int N, + int iK_UL, int Max_Secs, DP target_sumrule, bool refine); + + //void Scan_ODSLF (char whichDSF, DP Delta, DP N, int M, bool fixed_iK, int iKneeded, int Max_Secs, bool refine, int rank, int nr_processors); + void Scan_ODSLF (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int rank, int nr_processors); + void Scan_ODSLF (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, int Max_Secs, bool refine); + void Scan_ODSLF (char whichDSF, DP Delta, int N, int M, int iKneeded, int Max_Secs, bool refine); + void Scan_ODSLF (char whichDSF, DP Delta, int N, int M, int Max_Secs, bool refine); + + // Functions to prepare and wrapup parallel scans: + //void Prepare_Parallel_Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iK_UL, bool fixed_iK, int iKneeded, + void Prepare_Parallel_Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, + string defaultname, int paralevel, Vect rank_lower_paralevels, Vect nr_processors_lower_paralevels, + int nr_processors_at_newlevel); + //void Wrapup_Parallel_Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iK_UL, bool fixed_iK, int iKneeded, + void Wrapup_Parallel_Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, + string defaultname, int paralevel, Vect rank_lower_paralevels, Vect nr_processors_lower_paralevels, + int nr_processors_at_newlevel); + + void Prepare_Parallel_Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + int paralevel, Vect rank_lower_paralevels, Vect nr_processors_lower_paralevels, int nr_processors_at_newlevel); + void Wrapup_Parallel_Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + int paralevel, Vect rank_lower_paralevels, Vect nr_processors_lower_paralevels, int nr_processors_at_newlevel); + + + void Sort_RAW_File (const char ffsq_file[], char optionchar); + void Sort_RAW_File (const char ffsq_file[], char optionchar, char whichDSF); + + // Functions for data interpretation: + DP Smoothen_RAW_into_SF (string prefix, int iKmin, int iKmax, int DiK, + DP ommin, DP ommax, int Nom, DP gwidth, DP normalization, DP denom_sum_K); + DP Smoothen_RAW_into_SF (string prefix, Vect rawfilename, Vect weight, int iKmin, int iKmax, int DiK, + DP ommin, DP ommax, int Nom, DP gwidth, DP normalization, DP denom_sum_K); + //DP Smoothen_RAW_into_ASF (string prefix, int iKmin, int iKmax, DP ommin, DP ommax, int Nom, DP gwidth, + // DP normalization, DP denom_sum_K); // for autocorrelation. Deprecated during ABACUS++T_8, now done automatically within Smoothen_RAW_into_ASF. + void Write_K_File (DP Length, int iKmin, int iKmax); + void Write_Omega_File (int Nout_omega, DP omegamin, DP omegamax); + // Smoothen with gaussian width scaled with two-particle bandwidth + DP Smoothen_RAW_into_SF_LiebLin_Scaled (string prefix, DP L, int N, int iKmin, int iKmax, int DiK, DP ommin, DP ommax, int Nom, DP width, DP normalization); + + //**************************************************************************** + + struct Scan_Info { + + DP sumrule_obtained; + //long long int Nfull; + DP Nfull; // dimensionality of (sub)Hilbert space considered + long long int Ninadm; + long long int Ndata; + long long int Ndata_conv; + long long int Ndata_conv0; + //long long int CPU_ticks; + //long long int CPU_ticks_TOT; // including refine runs + double TT; // total computation time in seconds + + public: + Scan_Info(); // constructor, puts everything to zero + //Scan_Info (DP sr, long long int Nf, long long int Ni, long long int Nd, long long int Ndc, long long int Ndc0, long long int t); + Scan_Info (DP sr, DP Nf, long long int Ni, long long int Nd, long long int Ndc, long long int Ndc0, double t); + + void Save (const char* outfile_Cstr); + void Load (const char* infile_Cstr); + + inline Scan_Info& operator = (const Scan_Info& ref_info) + { + sumrule_obtained = ref_info.sumrule_obtained; + Nfull = ref_info.Nfull; + Ninadm = ref_info.Ninadm; + Ndata = ref_info.Ndata; + Ndata_conv = ref_info.Ndata_conv; + Ndata_conv0 = ref_info.Ndata_conv0; + //CPU_ticks = ref_info.CPU_ticks; + //CPU_ticks_TOT = ref_info.CPU_ticks_TOT; + TT = ref_info.TT; + + return(*this); + } + + inline Scan_Info& operator+= (const Scan_Info& ref_info) + { + if (this != &ref_info) { + sumrule_obtained += ref_info.sumrule_obtained; + Nfull += ref_info.Nfull; + Ninadm += ref_info.Ninadm; + Ndata += ref_info.Ndata; + Ndata_conv += ref_info.Ndata_conv; + Ndata_conv0 += ref_info.Ndata_conv0; + //CPU_ticks += ref_info.CPU_ticks; + //CPU_ticks_TOT += ref_info.CPU_ticks_TOT; + TT += ref_info.TT; + } + + return(*this); + } + + inline Scan_Info& operator-= (const Scan_Info& ref_info) + { + if (this != &ref_info) { + sumrule_obtained -= ref_info.sumrule_obtained; + Nfull -= ref_info.Nfull; + Ninadm -= ref_info.Ninadm; + Ndata -= ref_info.Ndata; + Ndata_conv -= ref_info.Ndata_conv; + Ndata_conv0 -= ref_info.Ndata_conv0; + //CPU_ticks -= ref_info.CPU_ticks; + //CPU_ticks_TOT -= ref_info.CPU_ticks_TOT; + TT -= ref_info.TT; + } + + return(*this); + } + + }; + + std::ostream& operator<< (std::ostream& s, const Scan_Info& info); + + template + Scan_Info General_Scan (char whichDSF, int iKmin, int iKmax, int iKmod, DP kBT, Tstate& AveragingState, Tstate& SeedScanState, + string defaultScanStatename, int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors); + + + //**************************************************************************** + + // Functions in src/SCAN/Descendents.cc: + /* + struct Descendent_Data { + + Vect label; + Vect type; + + Descendent_Data(); + + }; + */ + //Vect Descendents (const LiebLin_Bethe_State& ScanState, const LiebLin_Bethe_State& OriginState); + //Vect Descendents (const Heis_Bethe_State& ScanState, const Heis_Bethe_State& OriginState); + Vect Descendents (const LiebLin_Bethe_State& ScanState, const LiebLin_Bethe_State& OriginState, int type_required); + Vect Descendents (const Heis_Bethe_State& ScanState, const Heis_Bethe_State& OriginState, int type_required); + //Descendent_Data Descendents (const LiebLin_Bethe_State& ScanState, const LiebLin_Bethe_State& OriginState, int type_required); + //Descendent_Data Descendents (const Heis_Bethe_State& ScanState, const Heis_Bethe_State& OriginState, int type_required); + + //**************************************************************************** + /* DEPRECATED IN ABACUS++G_7 + struct Scan_Thread_List { + + int dim; // Dimension of vectors + int nthreads_tot; // Number of entries used; must be <= dim + int nthreads_done; // how many threads have already been processed + + ////Vect_DP omega; + //Vect omega; + //Vect_INT iK; + //Vect_DP abs_data_value; // Measure used to optimize scanning. This can be either the energy ('Z'), MEsq, omega * MEsq, ... + Vect abs_data_value; // Measure used to optimize scanning. This can be either the energy ('Z'), MEsq, omega * MEsq, ... + Vect label; + //Vect sector_lowest_raisable; + Vect type; // which type of descendent is needed: 0 == same nr of ph exc, 1 == one more ph exc. + Vect isdone; // whether the thread has been processed or not + + Scan_Thread_List (); // + Scan_Thread_List (int Nentries); + + bool Increase_Size (int nr_to_add); // resizes the vectors to accommodate up to nr_to_add additional entries + + //void Include_Thread (DP omega_ref, int iK_ref, DP abs_data_value_ref, string label_ref, int type_ref); + void Include_Thread (DP abs_data_value_ref, string label_ref, int type_ref); + void Order_in_abs_data_value (); // orders in decreasing abs_data_value, for correlation functions + //void Order_in_omega (); // orders in increasing omega, for partition function + //DP Highest_abs_data_value_ordered (DP omega_MIN, DP omega_MAX); // returns highest abs_data_value in window omega_MIN to omega_MAX, assuming list is ordered + //DP Highest_abs_data_value (DP omega_MIN, DP omega_MAX); // returns highest abs_data_value in window omega_MIN to omega_MAX + DP Highest_abs_data_value (); // returns highest abs_data_value + DP kth_highest_abs_data_value (int k); + bool Exists_data_value_greater_than (DP value); // whether at least one thread exceeds this value + //DP Lowest_omega (); // returns lowest omega + void Merge (const Scan_Thread_List& reflist); + //void Remove_Threads (int ithread_down, int ithread_up); // DEACTIVATED IN ABACUS++G_2 + //int Remove_Threads (const Vect& threads_done); // DEACTIVATED IN ABACUS++G_2 + void Remove_Done_Threads (); // removes done threads, replaces Remove_Threads (from ABACUS++G_2 onwards) + void Clear (); + void Save (const char* outfile_Cstr); + void Load (const char* thrfile_Cstr); + + inline Scan_Thread_List& operator = (const Scan_Thread_List& reflist) + { + dim = reflist.dim; + nthreads_tot = reflist.nthreads_tot; + nthreads_done = reflist.nthreads_done; + //omega = reflist.omega; + //iK = reflist.iK; + abs_data_value = reflist.abs_data_value; + label = reflist.label; + type = reflist.type; + isdone = reflist.isdone; + + return(*this); + } + + }; + */ + /* + struct Scan_Thread_Set { + + // By convention, a Scan_Thread_Set contains a list of threads which are yet to be descended. + + int nlists = 6400; // number of threads lists, fixed to this number by convention. + DP logscale = (1.0/64) * log(2.0); // each separate list contains threads differing by a scale factor of 2^{1/64} \approx 1.01. + // The number of lists thus covers abs_data_values from 1 down to about 10^-30. + // The list index of a thread is given by -log(data_value)/logscale, and forced between 0 and nlists -1. + Vect dim; + Vect nthreads_tot; + Vect nthreads_done; + + Vect > label; + Vect > type; // which type of descendent is needed + Vect > isdone; // whether the thread has been processed or not + + Scan_Thread_Set (); + + bool Increase_Size (int il, int nr_to_add); + void Include_Thread (DP abs_data_value_ref, string label_ref, int type_ref); + void Merge (const Scan_Thread_Set& refset); + void Remove_Done_Threads (int il); + void Remove_Done_Threads (); + void Clear (); + + void Save (const char* outfile_Cstr); + void Load (const char* thrfile_Cstr); + }; + */ + + + struct Scan_Thread { + + string label; + int type; + + Scan_Thread (); + Scan_Thread (string label_ref, int type_ref) { + label = label_ref; + type = type_ref; + } + + Scan_Thread& operator= (const Scan_Thread& RefThread); + }; + + + struct Scan_Thread_Data { + + // By convention, a Scan_Thread_Data object handles a list of threads which are yet to be descended. + // Improvement on Scan_Thread_Set used up to ABACUS++G_7, saving data to disk instead of holding it in memory. + + int nlists = 6400; // number of threads lists, fixed to this number by convention. + DP logscale = (1.0/64) * log(2.0); // each separate list contains threads differing by a scale factor of 2^{1/64} \approx 1.01. + //int nlists = 1600; // number of threads lists, fixed to this number by convention. + //DP logscale = (1.0/16) * log(2.0); // each separate list contains threads differing by a scale factor of 2^{1/64} \approx 1.01. + // The number of lists thus covers abs_data_values from 1 down to about 10^-30. + // The list index of a thread is given by -log(data_value)/logscale, and forced between 0 and nlists -1. + + string thrdir_name; // directory in which threads files are saved. + + Vect nthreads_total; + + Vect nthreads_on_disk; + int lowest_il_with_nthreads_neq_0; + + // In-memory storage, for adding threads efficiently without constantly writing to disk + // These objects are saved to disk when Next_Scan_Threads are called. + Vect dim; + Vect nthreads_in_memory; + Vect > label; + Vect > type; // which type of descendent is needed + + Vect filename; + //Vect file; + //Vect file_is_open; + + Scan_Thread_Data (); + Scan_Thread_Data (string thrdir_name_ref, bool refine); + ~Scan_Thread_Data (); + + bool Increase_Memory_Size (int il, int nr_to_add); + void Include_Thread (DP abs_data_value_ref, string label_ref, int type_ref); + void Include_Thread (int il, string label_ref, int type_ref); + + Vect Extract_Next_Scan_Threads (); // returns a vector of the threads that are next in line. By defn, all threads with index il == lowest_il_with_nthreads_neq_0. These are removed from the object. + Vect Extract_Next_Scan_Threads (int min_nr); // as above, but returns a minimum of min_nr threads. + + + void Flush_to_Disk (int il); + void Save (); + void Load (); + }; + + + + //**************************************************************************** + + // To populate a list of states for scanning: + + inline void Scan_for_Possible_Bases (const Vect SeedNrap, const Vect Str_L, + int Mdown_remaining, Vect& possible_base_label, int& nfound, int nexc_max_used, + int base_level_to_scan, Vect& Nrapidities) + { + if (Mdown_remaining < 0) { JSCerror("Scan_for_Possible_Bases: shouldn't be here..."); } // reached inconsistent point + + //cout << "Called Scan_for_Possible_Bases with Mdown_remaining = " << Mdown_remaining + //<< "\t" << possible_base_label << "\tnfound " << nfound << "\tnexc_max_used " << nexc_max_used + //<< "\tbase_level_to_scan " << base_level_to_scan << "\tNrap " << Nrapidities << endl; + + if (base_level_to_scan == 0) { + if (Str_L[0] != 1) JSCerror("Str_L[0] != 1 in JSC_Scan.h Scan_for_Possible_Bases."); + Nrapidities[0] = Mdown_remaining; + + // Set label: + stringstream M0out; + M0out << Nrapidities[0]; + possible_base_label[nfound] = M0out.str(); + for (int itype = 1; itype < Nrapidities.size(); ++itype) + if (Nrapidities[itype] > 0) { + possible_base_label[nfound] += TYPESEP; + stringstream typeout; + typeout << itype; + possible_base_label[nfound] += typeout.str(); + possible_base_label[nfound] += EXCSEP; + stringstream Mout; + Mout << Nrapidities[itype]; + possible_base_label[nfound] += Mout.str(); + } + nfound++; + } + + else { + // Preserve the number of strings at this level as compared to SeedState: + Nrapidities[base_level_to_scan] = SeedNrap[base_level_to_scan]; + if (Mdown_remaining - Str_L[base_level_to_scan] * Nrapidities[base_level_to_scan] >= 0) + Scan_for_Possible_Bases (SeedNrap, Str_L, Mdown_remaining - Str_L[base_level_to_scan] * Nrapidities[base_level_to_scan], + possible_base_label, nfound, nexc_max_used, base_level_to_scan - 1, Nrapidities); + + // Reduce number of strings at this level as compared to SeedState: + //for (int i = 1; i <= JSC::min(SeedNrap[base_level_to_scan], (Str_L[base_level_to_scan] == 0 ? 0 : nexc_max_used/Str_L[base_level_to_scan])); ++i) { + for (int i = 1; i <= JSC::min(SeedNrap[base_level_to_scan], nexc_max_used/Str_L[base_level_to_scan]); ++i) { + Nrapidities[base_level_to_scan] = SeedNrap[base_level_to_scan] - i; + if (Mdown_remaining - Str_L[base_level_to_scan] * Nrapidities[base_level_to_scan] >= 0) + Scan_for_Possible_Bases (SeedNrap, Str_L, Mdown_remaining - Str_L[base_level_to_scan] * Nrapidities[base_level_to_scan], + possible_base_label, nfound, nexc_max_used - i*Str_L[base_level_to_scan], base_level_to_scan - 1, Nrapidities); + } + // Increase the number of strings at this level as compared to SeedState: + for (int i = 1; i <= JSC::min(Mdown_remaining/Str_L[base_level_to_scan], nexc_max_used/Str_L[base_level_to_scan]); ++i) { + Nrapidities[base_level_to_scan] = SeedNrap[base_level_to_scan] + i; + if (Mdown_remaining - Str_L[base_level_to_scan] * Nrapidities[base_level_to_scan] >= 0) + Scan_for_Possible_Bases (SeedNrap, Str_L, Mdown_remaining - Str_L[base_level_to_scan] * Nrapidities[base_level_to_scan], + possible_base_label, nfound, nexc_max_used - i*Str_L[base_level_to_scan], base_level_to_scan - 1, Nrapidities); + } + } + } + + + inline Vect Possible_Bases (const Vect SeedNrap, const Vect Str_L, int Mdown)//const Heis_Bethe_State& SeedState) + { + //int nexc_max_used = JSC::min(NEXC_MAX_HEIS, 2*(Mdown/2)); // since each inner sector can contain at most N/2 holes. + int nexc_max_used = NEXC_MAX_HEIS; + + Vect possible_base_label (1000); + int nfound = 0; + Vect Nrapidities = SeedNrap; + int Mdown_remaining = Mdown; + + Scan_for_Possible_Bases (SeedNrap, Str_L, Mdown_remaining, possible_base_label, nfound, nexc_max_used, SeedNrap.size() - 1, Nrapidities); + + // Copy results into a clean vector: + Vect possible_base_label_found (nfound); + for (int i = 0; i < nfound; ++i) possible_base_label_found[i] = possible_base_label[i]; + + //cout << "In Possible_Bases: possible_base_label_found = " << possible_base_label_found << endl; + + return(possible_base_label_found); + } + + + + + + + + + //**************************************************************************** + + template + class Scan_State_List { + + public: + int ndef; + Vect State; + Vect base_label; + Vect info; // info for base and type of State[n] + //Vect descended; // true only if base has been descended + Vect flag_for_scan; // set to true, next round of scanning will use this base/type + Vect scan_attempted; // whether this has already been attempted + + public: + inline Scan_State_List (char whichDSF, const Tstate& SeedScanState); + + public: + inline Tstate& Return_State (string base_label_ref); // returns a state corresponding to same base and type + inline void Populate_List (char whichDSF, const Tstate& SeedScanState); // creates all types of states containing up to nexc_max excitations + inline void Include_Info (Scan_Info& info_to_add, string base_label_ref); + inline void Raise_Scanning_Flags (DP threshold); // checks whether base/type should be scanned based on simpler base/type combinations + + inline void Order_in_SRC (); + inline void Save_Info (const char* sumfile_Cstr); + inline void Load_Info (const char* sumfile_Cstr); + }; + + // Do the explicit class specializations: + + template<> + //inline Scan_State_List::Scan_State_List (char whichDSF, const LiebLin_Bethe_State& RefState) + inline Scan_State_List::Scan_State_List (char whichDSF, const LiebLin_Bethe_State& SeedScanState) + : ndef(0), State(Vect(MAX_STATE_LIST_SIZE)), base_label(Vect(MAX_STATE_LIST_SIZE)), + info(Vect(MAX_STATE_LIST_SIZE)), flag_for_scan(Vect(false, MAX_STATE_LIST_SIZE)), + scan_attempted(Vect(false, MAX_STATE_LIST_SIZE)) + { + //if (whichDSF == 'Z' || whichDSF == 'd' || whichDSF == 'q' || whichDSF == '1') + //State[0] = LiebLin_Bethe_State (RefState.c_int, RefState.L, RefState.N); + //else if (whichDSF == 'g') State[0] = LiebLin_Bethe_State (RefState.c_int, RefState.L, RefState.N + 1); + //else if (whichDSF == 'o') State[0] = LiebLin_Bethe_State (RefState.c_int, RefState.L, RefState.N - 1); + //else JSCerror("Unknown whichDSF in Scan_State_List + //inline Scan_State_List::Scan_State_List (char whichDSF, const XXZ_Bethe_State& RefState) + inline Scan_State_List::Scan_State_List (char whichDSF, const XXZ_Bethe_State& SeedScanState) + : ndef(0), State(Vect(MAX_STATE_LIST_SIZE)), base_label(Vect(MAX_STATE_LIST_SIZE)), + info(Vect(MAX_STATE_LIST_SIZE)), flag_for_scan(Vect(false, MAX_STATE_LIST_SIZE)), + scan_attempted(Vect(false, MAX_STATE_LIST_SIZE)) + { + //if (whichDSF == 'Z' || whichDSF == 'z') State[0] = XXZ_Bethe_State(RefState.chain, RefState.base.Mdown); + //else if (whichDSF == 'm') State[0] = XXZ_Bethe_State(RefState.chain, RefState.base.Mdown - 1); + //else if (whichDSF == 'p') State[0] = XXZ_Bethe_State(RefState.chain, RefState.base.Mdown + 1); + //else JSCerror("Unknown whichDSF in Scan_State_List + //inline Scan_State_List::Scan_State_List (char whichDSF, const XXX_Bethe_State& RefState) + inline Scan_State_List::Scan_State_List (char whichDSF, const XXX_Bethe_State& SeedScanState) + : ndef(0), State(Vect(MAX_STATE_LIST_SIZE)), base_label(Vect(MAX_STATE_LIST_SIZE)), + info(Vect(MAX_STATE_LIST_SIZE)), flag_for_scan(Vect(false, MAX_STATE_LIST_SIZE)), + scan_attempted(Vect(false, MAX_STATE_LIST_SIZE)) + { + //if (whichDSF == 'Z' || whichDSF == 'z' || whichDSF == 'a' || whichDSF == 'q') State[0] = XXX_Bethe_State(RefState.chain, RefState.base.Mdown); + //else if (whichDSF == 'm' || whichDSF == 'b') State[0] = XXX_Bethe_State(RefState.chain, RefState.base.Mdown - 1); + //else if (whichDSF == 'p') State[0] = XXX_Bethe_State(RefState.chain, RefState.base.Mdown + 1); + //else if (whichDSF == 'c') State[0] = XXX_Bethe_State(RefState.chain, RefState.base.Mdown - 2); + //else JSCerror("Unknown whichDSF in Scan_State_List + //inline Scan_State_List::Scan_State_List (char whichDSF, const XXZ_gpd_Bethe_State& RefState) + inline Scan_State_List::Scan_State_List (char whichDSF, const XXZ_gpd_Bethe_State& SeedScanState) + : ndef(0), State(Vect(MAX_STATE_LIST_SIZE)), base_label(Vect(MAX_STATE_LIST_SIZE)), + info(Vect(MAX_STATE_LIST_SIZE)), flag_for_scan(Vect(false, MAX_STATE_LIST_SIZE)), + scan_attempted(Vect(false, MAX_STATE_LIST_SIZE)) + { + //if (whichDSF == 'Z' || whichDSF == 'z') State[0] = XXZ_gpd_Bethe_State(RefState.chain, RefState.base.Mdown); + //else if (whichDSF == 'm') State[0] = XXZ_gpd_Bethe_State(RefState.chain, RefState.base.Mdown - 1); + //else if (whichDSF == 'p') State[0] = XXZ_gpd_Bethe_State(RefState.chain, RefState.base.Mdown + 1); + //else JSCerror("Unknown whichDSF in Scan_State_List + inline Scan_State_List::Scan_State_List (char whichDSF, const ODSLF_XXZ_Bethe_State& RefState) + : ndef(0), State(Vect(MAX_STATE_LIST_SIZE)), base_label(Vect(MAX_STATE_LIST_SIZE)), + info(Vect(MAX_STATE_LIST_SIZE)), flag_for_scan(Vect(false, MAX_STATE_LIST_SIZE)), + scan_attempted(Vect(false, MAX_STATE_LIST_SIZE)) + { + if (whichDSF == 'Z' || whichDSF == 'z') State[0] = ODSLF_XXZ_Bethe_State(RefState.chain, RefState.base.Mdown); + else if (whichDSF == 'm') State[0] = ODSLF_XXZ_Bethe_State(RefState.chain, RefState.base.Mdown - 1); + else if (whichDSF == 'p') State[0] = ODSLF_XXZ_Bethe_State(RefState.chain, RefState.base.Mdown + 1); + else JSCerror("Unknown whichDSF in Scan_State_List + inline LiebLin_Bethe_State& Scan_State_List::Return_State (string base_label_ref) + { + int n = 0; + while (n < ndef && base_label_ref.compare(base_label[n]) != 0) n++; + + if (n == ndef) { + State[n] = State[0]; + base_label[n] = base_label_ref; + info[n].Nfull = 1LL; // Nfull not definable for LiebLin + ndef++; + } + + return(State[n]); + } + + template<> + inline XXZ_Bethe_State& Scan_State_List::Return_State (string base_label_ref) + { + int n = 0; + while (n < ndef && base_label_ref.compare(base_label[n]) != 0) n++; + + if (n == ndef) { + Heis_Base checkbase (State[0].chain, base_label_ref); + State[n] = XXZ_Bethe_State (State[0].chain, checkbase); + info[n].Nfull = checkbase.dimH; + ndef++; + } + + return(State[n]); + } + + template<> + inline XXX_Bethe_State& Scan_State_List::Return_State (string base_label_ref) + { + int n = 0; + while (n < ndef && base_label_ref.compare(base_label[n]) != 0) n++; + + if (n == ndef) { + Heis_Base checkbase (State[0].chain, base_label_ref); + State[n] = XXX_Bethe_State (State[0].chain, checkbase); + info[n].Nfull = checkbase.dimH; + ndef++; + } + + return(State[n]); + } + + template<> + inline XXZ_gpd_Bethe_State& Scan_State_List::Return_State (string base_label_ref) + { + int n = 0; + while (n < ndef && base_label_ref.compare(base_label[n]) != 0) n++; + + if (n == ndef) { + Heis_Base checkbase (State[0].chain, base_label_ref); + State[n] = XXZ_gpd_Bethe_State (State[0].chain, checkbase); + info[n].Nfull = checkbase.dimH; + ndef++; + } + + return(State[n]); + } + + /* + template<> + inline ODSLF_XXZ_Bethe_State& Scan_State_List::Return_State (long long int base_id_ref, long long int type_id_ref) + { + int n = 0; + while (n < ndef && !(base_id_ref == State[n].base_id && type_id_ref == State[n].type_id)) n++; + + if (n == ndef) { + State[n] = ODSLF_XXZ_Bethe_State (State[0].chain, base_id_ref, type_id_ref); + info[n].Nfull = State[n].maxid + 1LL; + ndef++; + } + + return(State[n]); + } + */ + + template<> + //inline void Scan_State_List::Populate_List (char whichDSF, const LiebLin_Bethe_State& RefScanState) + inline void Scan_State_List::Populate_List (char whichDSF, const LiebLin_Bethe_State& SeedScanState) + { + // For LiebLin_Bethe_State: only one base is used, so there is only one state here. + + if (ndef != 0) JSCerror("Please only populate a virgin Scan_State_List."); + + stringstream baselabel; + baselabel << State[0].N; + base_label[0] = baselabel.str(); + stringstream label0; + label0 << State[0].N << LABELSEP << JSCcoding[0] << LABELSEP;//"_0_"; + //State[0].Set_to_Label (label0.str(), RefScanState.Ix2); + State[0].Set_to_Label (label0.str(), SeedScanState.Ix2); + info[0].Nfull = 1LL; // Nfull not definable for LiebLin + ndef = 1; + } + /* + template<> + //inline void Scan_State_List::Populate_List (char whichDSF, const XXZ_Bethe_State& RefScanState) + inline void Scan_State_List::Populate_List_pre_plusplusG_8 (char whichDSF, const XXZ_Bethe_State& SeedScanState) + { + // creates all types of states containing up to nexc_max excitations + + if (ndef != 0) JSCerror("Please only populate a virgin Scan_State_List."); + + //cout << "In Populate_List: " << State[0] << endl; + + //Vect bases_id = State[0].chain.Possible_Bases (State[0].base.Mdown); // returns a vector of possible bases + //cout << "Mdown = " << State[0].base.Mdown << "\tPossible bases size: " << bases_id.size() << "\tPossible bases: " << bases_id << endl; + Vect bases_label = State[0].chain.Possible_Bases (State[0].base.Mdown); // returns a vector of possible bases + //cout << "Mdown = " << State[0].base.Mdown << "\tPossible bases size: " << bases_label.size() << "\tPossible bases: " << bases_label << endl; + + + for (int ib = 0; ib < bases_label.size(); ++ib) { + + //cout << "ib = " << ib << "\tlabel " << bases_label[ib] << endl; + Heis_Base checkbase (State[0].chain, bases_label[ib]); + //cout << "Done building base for label " << bases_label[ib] << endl; + + State[ndef] = XXZ_Bethe_State (State[0].chain, checkbase); + base_label[ndef] = bases_label[ib]; + info[ndef].Nfull = 1LL; + ndef++; + if (ndef >= MAX_STATE_LIST_SIZE) JSCerror("Increase number of elements in ScanStateList."); + } + } + */ + + /* + // ++G_7 versions: + template<> + //inline void Scan_State_List::Populate_List (char whichDSF, const XXZ_Bethe_State& RefScanState) + inline void Scan_State_List::Populate_List (char whichDSF, const XXZ_Bethe_State& SeedScanState) + { + // creates all types of states containing up to nexc_max excitations + + if (ndef != 0) JSCerror("Please only populate a virgin Scan_State_List."); + + //cout << "In Populate_List: " << State[0] << endl; + + //Vect bases_id = State[0].chain.Possible_Bases (State[0].base.Mdown); // returns a vector of possible bases + //cout << "Mdown = " << State[0].base.Mdown << "\tPossible bases size: " << bases_id.size() << "\tPossible bases: " << bases_id << endl; + Vect bases_label = State[0].chain.Possible_Bases (State[0].base.Mdown); // returns a vector of possible bases + //cout << "Mdown = " << State[0].base.Mdown << "\tPossible bases size: " << bases_label.size() << "\tPossible bases: " << bases_label << endl; + + + for (int ib = 0; ib < bases_label.size(); ++ib) { + + //cout << "ib = " << ib << "\tlabel " << bases_label[ib] << endl; + Heis_Base checkbase (State[0].chain, bases_label[ib]); + //cout << "Done building base for label " << bases_label[ib] << endl; + + State[ndef] = XXZ_Bethe_State (State[0].chain, checkbase); + base_label[ndef] = bases_label[ib]; + info[ndef].Nfull = 1LL; + ndef++; + if (ndef >= MAX_STATE_LIST_SIZE) JSCerror("Increase number of elements in ScanStateList."); + } + } + + template<> + //inline void Scan_State_List::Populate_List (char whichDSF, const XXX_Bethe_State& RefScanState) + inline void Scan_State_List::Populate_List (char whichDSF, const XXX_Bethe_State& SeedScanState) + { + // creates all types of states containing up to nexc_max excitations + + if (ndef != 0) JSCerror("Please only populate a virgin Scan_State_List."); + + //cout << "In Populate_List: " << State[0] << endl; + + // To take infinite rapidities into account, we use intermediate states with up to 2 less finite rapidities (1 for Szz, 2 for Spm) + int nrinfrapmax = 0; + if (whichDSF == 'z') nrinfrapmax = 1; + else if (whichDSF == 'p') nrinfrapmax = JSC::min(2, SeedScanState.base.Mdown); + + //if (whichDSF == 'z' && RefScanState.chain.Nsites == 2*RefScanState.base.Mdown) + //if (whichDSF == 'z' && SeedScanState.chain.Nsites == 2*SeedScanState.base.Mdown) + //JSCerror("Szz at zero field for XXX is simply Smp/2. Compute the latter instead."); + //if (whichDSF == 'p' && RefScanState.chain.Nsites == 2*(RefScanState.base.Mdown + 1)) + //if (whichDSF == 'p' && SeedScanState.chain.Nsites/2 <= SeedScanState.base.Mdown) + //JSCerror("Spm for XXX at M >= N/2 is not implemented."); + + for (int nrinfrap = 0; nrinfrap <= nrinfrapmax; ++nrinfrap) { + + //Vect bases_id = State[0].chain.Possible_Bases (State[0].base.Mdown); // returns a vector of possible bases + Vect bases_label = State[0].chain.Possible_Bases (State[0].base.Mdown - nrinfrap); // returns a vector of possible bases + + //cout << "Mdown = " << State[0].base.Mdown << "\tPossible bases size: " << bases_id.size() << "\tPossible bases: " << bases_id << endl; + + for (int ib = 0; ib < bases_label.size(); ++ib) { + + Heis_Base checkbase (State[0].chain, bases_label[ib]); + + State[ndef] = XXX_Bethe_State (State[0].chain, checkbase); + base_label[ndef] = bases_label[ib]; + //cout << "nrinfrap = " << nrinfrap << "\tIncluding base " << base_label[ndef] << endl; + info[ndef].Nfull = 1LL; + ndef++; + if (ndef >= MAX_STATE_LIST_SIZE) JSCerror("Increase number of elements in ScanStateList."); + } + } // for nrinfrap + //cout << "Done populating." << endl; + } + + template<> + //inline void Scan_State_List::Populate_List (char whichDSF, const XXZ_gpd_Bethe_State& RefScanState) + inline void Scan_State_List::Populate_List (char whichDSF, const XXZ_gpd_Bethe_State& SeedScanState) + { + // creates all types of states containing up to nexc_max excitations + + if (ndef != 0) JSCerror("Please only populate a virgin Scan_State_List."); + + //cout << "In Populate_List: " << State[0] << endl; + + //Vect bases_id = State[0].chain.Possible_Bases (State[0].base.Mdown); // returns a vector of possible bases + Vect bases_label = State[0].chain.Possible_Bases (State[0].base.Mdown); // returns a vector of possible bases + + //cout << "Mdown = " << State[0].base.Mdown << "\tPossible bases size: " << bases_id.size() << "\tPossible bases: " << bases_id << endl; + + for (int ib = 0; ib < bases_label.size(); ++ib) { + + Heis_Base checkbase (State[0].chain, bases_label[ib]); + + State[ndef] = XXZ_gpd_Bethe_State (State[0].chain, checkbase); + base_label[ndef] = bases_label[ib]; + info[ndef].Nfull = 1LL; + ndef++; + if (ndef >= MAX_STATE_LIST_SIZE) JSCerror("Increase number of elements in ScanStateList."); + } + } + */ + + // ++G_8 versions: + template<> + //inline void Scan_State_List::Populate_List (char whichDSF, const XXZ_Bethe_State& RefScanState) + inline void Scan_State_List::Populate_List (char whichDSF, const XXZ_Bethe_State& SeedScanState) + { + // creates all types of states containing up to nexc_max excitations + + if (ndef != 0) JSCerror("Please only populate a virgin Scan_State_List."); + + // We assume that SeedScanState has quantum numbers which are set according to the relevant AveragingState. + + // This function creates a list of states with other bases in the vicinity of that of SeedScanState, + // matching the quantum numbers as closely as possible. + + Vect Str_L(SeedScanState.chain.Nstrings); + for (int i = 0; i < SeedScanState.chain.Nstrings; ++i) Str_L[i] = SeedScanState.chain.Str_L[i]; + + // First of all, we create a list of the possible bases themselves. + Vect bases_label = Possible_Bases (SeedScanState.base.Nrap, Str_L, SeedScanState.base.Mdown); // returns a vector of possible bases + + for (int ib = 0; ib < bases_label.size(); ++ib) { + + //cout << "ib = " << ib << "\tlabel " << bases_label[ib] << endl; + Heis_Base checkbase (State[0].chain, bases_label[ib]); + //cout << "Done building base for label " << bases_label[ib] << endl; + + State[ndef] = XXZ_Bethe_State (State[0].chain, checkbase); + State[ndef].Set_to_Closest_Matching_Ix2_fixed_Base (SeedScanState); + State[ndef].Set_Label_from_Ix2 (State[ndef].Ix2); // sets to trivial label for this base + //cout << "\tSet state label to " << State[ndef].label << endl; + //cout << "\tSet state to " << State[ndef] << endl; + base_label[ndef] = bases_label[ib]; + info[ndef].Nfull = State[ndef].base.dimH; + ndef++; + if (ndef >= MAX_STATE_LIST_SIZE) JSCerror("Increase number of elements in ScanStateList."); + } + } + + template<> + //inline void Scan_State_List::Populate_List (char whichDSF, const XXX_Bethe_State& RefScanState) + inline void Scan_State_List::Populate_List (char whichDSF, const XXX_Bethe_State& SeedScanState) + { + // creates all types of states containing up to nexc_max excitations + + if (ndef != 0) JSCerror("Please only populate a virgin Scan_State_List."); + + // We assume that SeedScanState has quantum numbers which are set according to the relevant AveragingState. + + // This function creates a list of states with other bases in the vicinity of that of SeedScanState, + // matching the quantum numbers as closely as possible. + + Vect Str_L(SeedScanState.chain.Nstrings); + for (int i = 0; i < SeedScanState.chain.Nstrings; ++i) Str_L[i] = SeedScanState.chain.Str_L[i]; + + // To take infinite rapidities into account, we use intermediate states with up to 2 less finite rapidities (1 for Szz, 2 for Spm) + int nrinfrapmax = 0; + if (whichDSF == 'z') nrinfrapmax = 1; + else if (whichDSF == 'p') nrinfrapmax = JSC::min(2, SeedScanState.base.Mdown); + + //if (whichDSF == 'z' && RefScanState.chain.Nsites == 2*RefScanState.base.Mdown) + //if (whichDSF == 'z' && SeedScanState.chain.Nsites == 2*SeedScanState.base.Mdown) + //JSCerror("Szz at zero field for XXX is simply Smp/2. Compute the latter instead."); + //if (whichDSF == 'p' && RefScanState.chain.Nsites == 2*(RefScanState.base.Mdown + 1)) + //if (whichDSF == 'p' && SeedScanState.chain.Nsites/2 <= SeedScanState.base.Mdown) + //JSCerror("Spm for XXX at M >= N/2 is not implemented."); + + Vect Nrapmod = SeedScanState.base.Nrap; + + for (int nrinfrap = 0; nrinfrap <= nrinfrapmax; ++nrinfrap) { + + Nrapmod[0] = SeedScanState.base.Nrap[0] - nrinfrap; + if (Nrapmod[0] < 0) JSCerror("Putting too many rapidities at infinity in JSC_Scan.h: Possible_Bases."); + + Vect bases_label = Possible_Bases (Nrapmod, Str_L, SeedScanState.base.Mdown-nrinfrap); // returns a vector of possible bases + + for (int ib = 0; ib < bases_label.size(); ++ib) { + + Heis_Base checkbase (State[0].chain, bases_label[ib]); + + State[ndef] = XXX_Bethe_State (State[0].chain, checkbase); + State[ndef].Set_to_Closest_Matching_Ix2_fixed_Base (SeedScanState); + //State[ndef].Set_Label_from_Ix2 (State[ndef].Ix2); // sets to trivial label for this base + base_label[ndef] = bases_label[ib]; + //cout << "nrinfrap = " << nrinfrap << "\tIncluding base " << base_label[ndef] << endl; + info[ndef].Nfull = State[ndef].base.dimH; + ndef++; + if (ndef >= MAX_STATE_LIST_SIZE) JSCerror("Increase number of elements in ScanStateList."); + } + } // for nrinfrap + //cout << "Done populating." << endl; + } + + template<> + //inline void Scan_State_List::Populate_List (char whichDSF, const XXZ_gpd_Bethe_State& RefScanState) + inline void Scan_State_List::Populate_List (char whichDSF, const XXZ_gpd_Bethe_State& SeedScanState) + { + // creates all types of states containing up to nexc_max excitations + + if (ndef != 0) JSCerror("Please only populate a virgin Scan_State_List."); + + // We assume that SeedScanState has quantum numbers which are set according to the relevant AveragingState. + + // This function creates a list of states with other bases in the vicinity of that of SeedScanState, + // matching the quantum numbers as closely as possible. + + Vect Str_L(SeedScanState.chain.Nstrings); + for (int i = 0; i < SeedScanState.chain.Nstrings; ++i) Str_L[i] = SeedScanState.chain.Str_L[i]; + + // First of all, we create a list of the possible bases themselves. + Vect bases_label = Possible_Bases (SeedScanState.base.Nrap, Str_L, SeedScanState.base.Mdown); // returns a vector of possible bases + + for (int ib = 0; ib < bases_label.size(); ++ib) { + + Heis_Base checkbase (State[0].chain, bases_label[ib]); + + State[ndef] = XXZ_gpd_Bethe_State (State[0].chain, checkbase); + State[ndef].Set_to_Closest_Matching_Ix2_fixed_Base (SeedScanState); + //State[ndef].Set_Label_from_Ix2 (State[ndef].Ix2); // sets to trivial label for this base + base_label[ndef] = bases_label[ib]; + info[ndef].Nfull = State[ndef].base.dimH; + ndef++; + if (ndef >= MAX_STATE_LIST_SIZE) JSCerror("Increase number of elements in ScanStateList."); + } + } + + + /* + template<> + inline void Scan_State_List::Populate_List () + { + // creates all types of states containing up to nexc_max excitations + + if (ndef != 0) JSCerror("Please only populate a virgin Scan_State_List."); + + //cout << "In Populate_List: " << State[0] << endl; + + Vect bases_id = State[0].chain.Possible_Bases (State[0].base.Mdown); // returns a vector of possible bases + + //cout << "Mdown = " << State[0].base.Mdown << "\tPossible bases size: " << bases_id.size() << "\tPossible bases: " << bases_id << endl; + + for (int ib = 0; ib < bases_id.size(); ++ib) { + ODSLF_Base checkbase (State[0].chain, bases_id[ib]); + + Vect types_id = checkbase.Possible_Types (); // returns a vector of possible types + + //cout << "For base_id " << bases_id[ib] << "\t found types " << types_id << endl; + + for (int it = 0; it < types_id.size(); ++it) { + + if (bases_id[ib] < 1000000) { // FUDGE: consider only one-strings + //cout << "Populate list: constructing state: " << bases_id[ib] << "\t" << types_id[it] << endl; + State[ndef] = ODSLF_XXZ_Bethe_State (State[0].chain, bases_id[ib], types_id[it]); + //cout << "Populate list: before setting id: " << endl << State[ndef] << endl; + State[ndef].Set_to_id(0LL); + //cout << "Populate list: after setting id: " << endl << State[ndef] << endl; + info[ndef].Nfull = State[ndef].maxid + 1LL; + ndef++; + if (ndef >= MAX_STATE_LIST_SIZE) JSCerror("Increase number of elements in ScanStateList."); + } + } + + } + } + + */ + + + + template + inline void Scan_State_List::Include_Info (Scan_Info& info_to_add, string base_label_ref) + { + + int n = 0; + while (n < ndef && base_label_ref.compare(base_label[n]) != 0) n++; + + if (n == ndef) { + cout << "ndef = " << ndef << endl; + for (int i = 0; i < ndef; ++i) cout << base_label[i] << "\t"; + cout << endl; + cout << "base_label_ref " << base_label_ref << endl; + JSCerror("Did not find base_label_ref in Scan_State_List::Include_Info."); + } + + info[n] += info_to_add; + return; + } + + + /* + inline int Base_Distance (string base_label_1, string base_label_2) + { + int dist = 0; + + State_Label_Data labeldata1 = Extract_base_label_data (base_label_1); + State_Label_Data labeldata2 = Extract_base_label_data (base_label_2); + + return(dist); + } + */ + /* + inline int Type_Nr_Excitations (string base_label) + { + State_Label_Data labeldata = Extract_base_label_data (base_label); + + return(labeldata.nexc[0]); + } + */ + template + inline void Scan_State_List::Raise_Scanning_Flags (DP threshold) + { + flag_for_scan = true; + } + /* + { + // Reset whole flags vector: + flag_for_scan = false; + + // We first flag all base/types having up to 1 higher string and up to 2 excitations in sectors 0-3 + + for (int n = 0; n < ndef; ++n) + //if (info[n].Nfull == 0LL && Base_Distance (State[n].base_id, 0LL) <= 1 + if (!scan_attempted[n] && Base_Distance (base_label[n], base_label[0]) <= 1 + && Type_Nr_Excitations(base_label[n]) <= 3) + flag_for_scan[n] = true; + + // Second, if one base/type combination has given good results, + // we flag as yet unopened but slightly more complex base/types: + + //cout << "Called Raise_Scanning_Flags: ndef = " << ndef << endl; + + for (int n = 0; n < ndef; ++n) { + //cout << "Base/type " << State[n].base_id << "\t" << State[n].type_id << "\t" << info[n].sumrule_obtained << "\t" << threshold << endl; + if (info[n].Nfull > 0LL && info[n].sumrule_obtained > threshold) { + //cout << "Base/type " << State[n].base_id << "\t" << State[n].type_id << "\tflagging descendents:" << endl; + for (int n2 = 0; n2 < ndef; ++n2) { + //cout << "\tBase/type " << State[n2].base_id << "\t" << State[n2].type_id + // << "\t" << !scan_attempted[n2] << "\t" << (Base_Distance (State[n].base_id, State[n2].base_id) <= 1) + // << "\t" << (Type_Distance (State[n].type_id, State[n2].type_id) <= 2) << endl; + //if (info[n2].Nfull == 0LL && Base_Distance (State[n].base_id, State[n2].base_id) <= 1 + if (!scan_attempted[n2] && Base_Distance (base_label[n], base_label[n2]) <= 1 + && Type_Distance (base_label[n], base_label[n2]) <= 2) { + flag_for_scan[n2] = true; + //cout << "Flagging base/type " << State[n2].base_id << "\t" << State[n2].type_id + // << "\tfrom " << State[n].base_id << "\t" << State[n].type_id + // << "\tBase_Distance " << Base_Distance (State[n].base_id, State[n2].base_id) << "\tType_Distance " << Type_Distance (State[n].type_id, State[n2].type_id) << endl; + } + } + } + } + } + */ + template + inline void Scan_State_List::Order_in_SRC () + { + if (ndef > 0) { + + Vect_INT index(ndef); + for (int i = 0; i < ndef; ++i) index[i] = i; + + Vect sr (ndef); + for (int i = 0; i < ndef; ++i) sr[i] = info[i].sumrule_obtained; + + sr.QuickSort(index, 0, ndef - 1); + + Vect State_ordered(ndef); + Vect base_label_ordered(ndef); + Vect info_ordered(ndef); + Vect flag_for_scan_ordered(ndef); + + // Put data in proper order + for (int i = 0; i < ndef; ++i) { + State_ordered[i] = State[index[ndef - 1 - i] ]; + base_label_ordered[i] = base_label[index[ndef - 1 - i] ]; + info_ordered[i] = info[index[ndef - 1 - i] ]; + flag_for_scan_ordered[i] = flag_for_scan[index[ndef - 1 - i] ]; + } + + // Put back in *this object: + for (int i = 0; i < ndef; ++i) { + State[i] = State_ordered[i]; + base_label[i] = base_label_ordered[i]; + info[i] = info_ordered[i]; + flag_for_scan[i] = flag_for_scan_ordered[i]; + } // The rest are all simply 0. + } + // Done ! + } + + template + inline void Scan_State_List::Save_Info (const char* sumfile_Cstr) + { + ofstream outfile; + + outfile.open(sumfile_Cstr); + if (outfile.fail()) JSCerror("Could not open outfile... "); + + outfile.setf(ios::fixed); + //outfile.setf(ios::showpoint); + outfile.precision(16); + + //outfile << "base\t\tsumrule_obtained\tNfull\tNinadm\tNdata\tconv\tconv0\tT\tTT."; + //outfile << setw(20) << "base" << setw(25) << "sumrule_obtained" << setw(25) << "Nfull" << setw(10) << "Ninadm" << setw(10) << "Ndata" << setw(10) << "conv" << setw(10) << "conv0" << setw(10) << "T" << setw(10) << "TT."; + outfile << setw(20) << "base" << setw(25) << "sumrule_obtained" << setw(25) << "Nfull" << setw(10) << "Ninadm" << setw(10) << "Ndata" << setw(10) << "conv" << setw(10) << "conv0" << setw(10) << "TT."; + + for (int i = 0; i < ndef; ++i) + if (info[i].Nfull > 0.0) { + int TT_hr = int(info[i].TT/3600); + int TT_min = int((info[i].TT - 3600.0*TT_hr)/60); + //outfile << endl << base_label[i] << "\t" << info[i].sumrule_obtained + // << "\t" << info[i].Nfull << "\t" << info[i].Ninadm << "\t" << info[i].Ndata << "\t" << info[i].Ndata_conv << "\t" + // << info[i].Ndata_conv0 << "\t" << info[i].CPU_ticks/CLOCKS_PER_SEC << "\t" << info[i].CPU_ticks_TOT/CLOCKS_PER_SEC; + outfile << endl << setw(20) << base_label[i] << setw(25) << std::fixed << setprecision(20) << info[i].sumrule_obtained; + if (info[i].Nfull < 1.0e+10) outfile << setw(25) << std::fixed << setprecision(0) << info[i].Nfull; + else outfile << setw(25) << std::scientific << setprecision(16) << info[i].Nfull; + //outfile << setw(10) << info[i].Ninadm << setw(10) << info[i].Ndata << setw(10) << info[i].Ndata_conv << setw(10) << info[i].Ndata_conv0 << setw(10) << info[i].CPU_ticks/CLOCKS_PER_SEC << setw(10) << info[i].CPU_ticks_TOT/CLOCKS_PER_SEC; + //outfile << setw(10) << info[i].Ninadm << setw(10) << info[i].Ndata << setw(10) << info[i].Ndata_conv << setw(10) << info[i].Ndata_conv0 << setw(16) << std::fixed << std::showpoint << setprecision(3) << info[i].TT; + outfile << setw(10) << info[i].Ninadm << setw(10) << info[i].Ndata << setw(10) << info[i].Ndata_conv << setw(10) << info[i].Ndata_conv0 << setw(10) << TT_hr << " h " << TT_min << " m " << std::fixed << std::showpoint << setprecision(3) << info[i].TT - 3600.0*TT_hr - 60.0*TT_min << " s"; + } + outfile.close(); + } + + template + inline void Scan_State_List::Load_Info (const char* sumfile_Cstr) + { + ifstream infile; + infile.open(sumfile_Cstr); + if(infile.fail()) { + cout << endl << sumfile_Cstr << endl; + JSCerror("Could not open input file in Scan_State_List::Load_Info."); + } + + // Load first line, containing informative text: + char junk[256]; + infile.getline(junk, 256); + + // Now load the previous info's: + string base_label_ref; + //DP sr_ref, CPU_ref, CPU_TOT_ref; + DP sr_ref; + //long long int Nfull_ref, Ninadm_ref, Ndata_ref, conv_ref, conv0_ref; + DP Nfull_ref; + long long int Ninadm_ref, Ndata_ref, conv_ref, conv0_ref; + DP TT_ref; + int TT_hr, TT_min; + DP TT_sec; + char a; + + while (infile.peek() != EOF) { + //infile >> base_label_ref >> sr_ref >> Nfull_ref >> Ninadm_ref >> Ndata_ref >> conv_ref >> conv0_ref >> CPU_ref >> CPU_TOT_ref; + //infile >> base_label_ref >> sr_ref >> Nfull_ref >> Ninadm_ref >> Ndata_ref >> conv_ref >> conv0_ref >> TT_ref; + infile >> base_label_ref >> sr_ref >> Nfull_ref >> Ninadm_ref >> Ndata_ref >> conv_ref >> conv0_ref >> TT_hr >> a >> TT_min >> a >> TT_sec >> a; + TT_ref = 3600.0 * TT_hr + 60.0* TT_min + TT_sec; + //Scan_Info info_ref (sr_ref, Nfull_ref, Ninadm_ref, Ndata_ref, conv_ref, conv0_ref, CPU_ref); + Scan_Info info_ref (sr_ref, Nfull_ref, Ninadm_ref, Ndata_ref, conv_ref, conv0_ref, TT_ref); + (*this).Include_Info (info_ref, base_label_ref); + } + + infile.close(); + + return; + } + + +} // namespace JSC + +#endif diff --git a/include/JSC_Spec_Fns.h b/include/JSC_Spec_Fns.h new file mode 100644 index 0000000..2770d50 --- /dev/null +++ b/include/JSC_Spec_Fns.h @@ -0,0 +1,207 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_Spec_Fns.h + +Purpose: Defines special math functions. + +***********************************************************/ + +#ifndef _JSC_SPEC_FNS_H_ +#define _JSC_SPEC_FNS_H_ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + inline DP Cosine_Integral (DP x) + { + + // Returns the Cosine integral -\int_x^\infty dt \frac{\cos t}{t} + // Refer to GR[6] 8.23 + + if (x <= 0.0) { + cout << "Cosine_Integral called with real argument " << x << " <= 0, which is ill-defined because of the branch cut." << endl; + JSCerror(""); + } + + else if (x < 15.0) { // Use power series expansion + + // Ci (x) = gamma + \ln x + \sum_{n=1}^\infty (-1)^n x^{2n}/(2n (2n)!). + + int n = 1; + DP minonetothen = -1.0; + DP logxtothetwon = 2.0 * log(x); + DP twologx = 2.0 * log(x); + DP logtwonfact = log(2.0); + + DP series = minonetothen * exp(logxtothetwon - log(2.0 * n) - logtwonfact); + DP term_n; + + do { + n += 1; + minonetothen *= -1.0; + logxtothetwon += twologx; + logtwonfact += log((2.0 * n - 1.0) * 2.0 * n); + term_n = minonetothen * exp(logxtothetwon - log(2.0 * n) - logtwonfact); + series += term_n; + + } while (fabs(term_n) > 1.0e-16); + + + + + /* + // For improved convergence we pair terms up, DOESN'T WORK WELL + + // Ci (x) = gamma + \ln x - \sum{n = 1, 3, 5, ...} \frac{x^{2n}}{2n (2n)!} ( 1 - \frac{n}{n+1} \frac{x^2}{(2n+1)(2n+2)} ) + + int n = 1; + DP logxtothetwon = 2.0 * log(x); + DP logtwon = log(2.0); + DP logtwonfact = log(2.0); + DP xsq = x*x; + + DP series = exp(logxtothetwon - logtwon - logtwonfact) * (1 - xsq/((2.0 * n + 1.0) * (2.0 * n + 2.0) * (1.0 + 1.0/n))); + DP term_n; + DP twologx = 2.0 * log(x); + + do { + n += 2; + logxtothetwon += twologx; + logtwonfact += log((2.0 * n - 1.0) * 2.0 * n); + term_n = exp(logxtothetwon - log(2.0 * n) - logtwonfact) * (1 - xsq/((2.0 * n + 1.0) * (2.0 * n + 2.0) * (1.0 + 1.0/n)));; + series += term_n; + + } while (fabs(term_n) > 1.0e-16); + */ + + return(Euler_Mascheroni + log(x) + series); + } + + + else { // Use high x power series + + // Ci (x) = \frac{\sin x}{x} \sum_{n=0}^\infty (-1)^n (2n)! x^{-2n} - \frac{\cos x}{x} \sum_{n=0}^\infty (-1)^n (2n+1)! x^{-2n-1} + + int n = 0; + DP minonetothen = 1.0; + DP logxtothetwon = 0.0; + DP logxtothetwonplus1 = log(x); + DP twologx = 2.0 * log(x); + DP logtwonfact = 0.0; + DP logtwonplus1fact = 0.0; + + DP series1 = minonetothen * exp(logtwonfact - logxtothetwon); + DP series2 = minonetothen * exp(logtwonplus1fact - logxtothetwonplus1); + + do { + n += 1; + minonetothen *= -1.0; + logxtothetwon += twologx; + logxtothetwonplus1 += twologx; + logtwonfact += log((2.0 * n - 1.0) * 2.0 * n); + logtwonplus1fact += log(2.0 * n * (2.0 * n + 1)); + + series1 += minonetothen * exp(logtwonfact - logxtothetwon); + series2 += minonetothen * exp(logtwonplus1fact - logxtothetwonplus1); + + } while (n < 12); + + return((sin(x)/x) * series1 - (cos(x)/x) * series2); + + } + + + return(log(-1.0)); + } + + + /*********** Jacobi Theta functions *********/ + + inline DP Jacobi_Theta_1_q (DP u, DP q) { + + // Uses the summation formula. + // theta_1 (x) = 2 \sum_{n=1}^\infty (-1)^{n+1} q^{(n-1/2)^2} \sin (2n-1)u + // in which q is the nome. (GR 8.180.1) + // We always evaluate to numerical accuracy. + + if (q >= 1.0) JSCerror("Jacobi_Theta_1_q function called with q > 1."); + + + DP answer = 0.0; + DP contrib = 0.0; + DP qtonminhalfsq = pow(q, 0.25); // this will be q^{(n-1/2)^2} + DP qtotwon = pow(q, 2.0); // this will be q^{2n} + DP qsq = q*q; + int n = 1; + + do { + contrib = (n % 2 ? 2.0 : -2.0) * qtonminhalfsq * sin((2.0*n - 1.0)*u); + answer += contrib; + qtonminhalfsq *= qtotwon; + qtotwon *= qsq; + n++; + //cout << "\t\tn = " << n << "\tanswer = " << answer << "\tcontrib = " << contrib << "\tqtonminhalfsq = " << qtonminhalfsq << "\tqtotwon = " << qtotwon << endl; + } while (fabs(contrib/answer) > MACHINE_EPS); + + //cout << "\t\tJacobi_Theta_1: used " << n << " iterations." << "\tanswer = " << answer << "\tcontrib = " << contrib << "\tqtonminhalfsq = " << qtonminhalfsq << "\tqtotwon = " << qtotwon << endl; + return(answer); + } + + inline complex ln_Jacobi_Theta_1_q (complex u, complex q) { + + // This uses the product representation + // \theta_1 (x) = 2 q^{1/4} \sin{u} \prod_{n=1}^\infty (1 - 2 q^{2n} \cos 2u + q^{4n}) (1 - q^{2n}) + // (GR 8.181.2) + + complex contrib = 0.0; + complex qtotwon = q*q; // this will be q^{2n} + complex qsq = q*q; + complex twocos2u = 2.0 * cos(2.0*u); + int n = 1; + complex answer = log(2.0 * sin(u)) + 0.25 * log(q); + + do { + contrib = log((1.0 - twocos2u * qtotwon + qtotwon * qtotwon) * (1.0 - qtotwon)); + answer += contrib; + qtotwon *= qsq; + n++; + } while (abs(contrib) > 1.0e-12); + + return(answer); + } + + + /************ Barnes function ************/ + + inline DP ln_Gamma_for_Barnes_G_RE (Vect_DP args) + { + return(real(ln_Gamma(complex(args[0])))); + } + + inline DP ln_Barnes_G_RE (DP z) + { + // Implementation according to equation (28) of 2004_Adamchik_CPC_157 + // Restricted to real arguments. + + Vect_DP args (0.0, 2); + + DP req_rel_prec = 1.0e-6; + DP req_abs_prec = 1.0e-6; + int max_nr_pts = 10000; + Integral_result integ_ln_Gamma = Integrate_optimal (ln_Gamma_for_Barnes_G_RE, args, 0, 0.0, z - 1.0, req_rel_prec, req_abs_prec, max_nr_pts); + + return(0.5 * (z - 1.0) * (2.0 - z + logtwoPI) + (z - 1.0) * real(ln_Gamma(complex(z - 1.0))) - integ_ln_Gamma.integ_est); + } + +} // namespace JSC + +#endif // _JS_SPEC_FNS_H_ diff --git a/include/JSC_State_Ensemble.h b/include/JSC_State_Ensemble.h new file mode 100644 index 0000000..fd25f04 --- /dev/null +++ b/include/JSC_State_Ensemble.h @@ -0,0 +1,47 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_State_Ensemble.h + +Purpose: Define state ensembles. + + +***********************************************************/ + +#ifndef _ENS_ +#define _ENS_ + +#include "JSC.h" + +namespace JSC { + + + struct LiebLin_Diagonal_State_Ensemble { + + int nstates; + Vect state; + Vect weight; + + LiebLin_Diagonal_State_Ensemble (); + LiebLin_Diagonal_State_Ensemble (const LiebLin_Bethe_State& RefState, int nstates_req); + //LiebLin_Diagonal_State_Ensemble (const LiebLin_Bethe_State& RefState, int nstates_req, const Vect& weight_ref); + //LiebLin_Diagonal_State_Ensemble (DP c_int, DP L, int N, const Root_Density& rho, int nstates_req); + LiebLin_Diagonal_State_Ensemble (DP c_int, DP L, int N, const Root_Density& rho); + + LiebLin_Diagonal_State_Ensemble& operator= (const LiebLin_Diagonal_State_Ensemble& rhs); + void Load (DP c_int, DP L, int N, const char* ensfile_Cstr); + void Save (const char* ensfile_Cstr); + }; + + //LiebLin_Diagonal_State_Ensemble LiebLin_Thermal_Saddle_Point_Ensemble (DP c_int, DP L, int N, DP kBT, int nstates_req); + LiebLin_Diagonal_State_Ensemble LiebLin_Thermal_Saddle_Point_Ensemble (DP c_int, DP L, int N, DP kBT); + + +} // namespace JSC + +#endif diff --git a/include/JSC_TBA.h b/include/JSC_TBA.h new file mode 100644 index 0000000..43f0f53 --- /dev/null +++ b/include/JSC_TBA.h @@ -0,0 +1,147 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS. library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_TBA.h + +Purpose: Thermodynamic Bethe Ansatz general functions + + +***********************************************************/ + +#ifndef _TBA_ +#define _TBA_ + +#include "JSC.h" + +namespace JSC { + + struct Root_Density { + + int Npts; // how many points are used to describe each function + DP lambdamax; // what the largest rapidity is + Vect_DP lambda; // rapidity vector + Vect_DP dlambda; // differential element + Vect_DP value; // the root density itself + Vect_DP prev_value; // results of previous iteration + DP diff; // relative differences with previous iteration + bool value_infty_set; // boolean, true if asymptotic value set + DP value_infty; // asymptotic value, computed analytically + + Root_Density (); + Root_Density (int Npts_ref, DP lambdamax_ref); + + Root_Density& operator= (const Root_Density& RefDensity); + + void Save (const char* outfile_Cstr); + + DP Return_Value (DP lambda_ref); // evaluates the function for any argument using linear interpolation + void Set_Asymptotics (DP value_infty_ref); // sets value for lambda >= lambdamax + + Root_Density Compress_and_Match_Densities (DP comp_factor); // returns a Root_Density with fewer points + }; + + struct Root_Density_Set { + + int ntypes; + Vect epsilon; + int Npts_total; // sum of all Npts of epsilon's + DP diff; // sum of diff's of the epsilon's + + Root_Density_Set (); + Root_Density_Set (int ntypes_ref, int Npts_ref, DP lambdamax_ref); + Root_Density_Set (int ntypes_ref, Vect_INT Npts_ref, Vect_DP lambdamax_ref); + + Root_Density_Set& operator= (const Root_Density_Set& RefSet); + + void Insert_new_function (DP asymptotic_value); + void Extend_limits (Vect need_to_extend_limit); + void Insert_new_points (Vect > need_new_point_around); + + DP Return_Value (int n_ref, DP lambda_ref); // returns a value, no matter what. + + Root_Density_Set Return_Compressed_and_Matched_Set (DP comp_factor); + void Match_Densities (Root_Density_Set& RefSet); + + void Save (const char* outfile_Cstr); + }; + + + struct LiebLin_TBA_Solution { + + DP c_int; + DP mu; + DP kBT; + DP nbar; + DP ebar; // mean energy, \int d\lambda \lambda^2 \rho (\lambda) + DP sbar; // entropy per unit length + Root_Density epsilon; + Root_Density depsilon_dmu; + Root_Density rho; + Root_Density rhoh; + + LiebLin_TBA_Solution (DP c_int_ref, DP mu_ref, DP kBT_ref, DP req_diff, int Max_Secs); + }; + + // Functions defined in TBA_LiebLin.cc + Root_Density LiebLin_rho_GS (DP c_int, DP k_F, DP lambdamax, int Npts, DP req_prec); + DP Density_GS (Root_Density& rho_GS); + DP k_F_given_n (DP c_int, DP n, DP lambdamax, int Npts, DP req_prec); + Root_Density LiebLin_Z_GS (DP c_int, DP k_F, DP lambdamax, int Npts, DP req_prec); + Root_Density LiebLin_Fbackflow_GS (DP c_int, DP k_F, DP lambdamax, DP lambda, int Npts, DP req_prec); + // epsilon for a given mu: + Root_Density LiebLin_epsilon_TBA (DP c_int, DP mu, DP kBT, DP req_diff, int Max_Secs); + // depsilon/dmu for a given mu: + Root_Density LiebLin_depsilon_dmu_TBA (DP c_int, DP mu, DP kBT, DP req_diff, int Max_Secs, const Root_Density& epsilon); + Root_Density LiebLin_rho_TBA (DP kBT, const Root_Density& epsilon, const Root_Density& depsilon_dmu); + Root_Density LiebLin_rhoh_TBA (DP kBT, const Root_Density& epsilon, const Root_Density& depsilon_dmu); + DP LiebLin_nbar_TBA (const Root_Density& rho); + DP LiebLin_ebar_TBA (const Root_Density& rho); + DP LiebLin_sbar_TBA (const Root_Density& rho, const Root_Density& rhoh); + LiebLin_TBA_Solution LiebLin_TBA_Solution_fixed_nbar (DP c_int, DP nbar_required, DP kBT, DP req_diff, int Max_Secs); + LiebLin_TBA_Solution LiebLin_TBA_Solution_fixed_nbar_ebar (DP c_int, DP nbar_required, DP ebar_required, DP req_diff, int Max_Secs); + LiebLin_Bethe_State Discretized_LiebLin_Bethe_State (DP c_int, DP L, int N, const Root_Density& rho); + + // Functions defined in TBA_XXZ.cc + DP XXZ_phi1_kernel (DP zeta, DP lambda); + DP XXZ_phi2_kernel (DP zeta, DP lambda); + DP XXZ_a1_kernel (DP sinzeta, DP coszeta, DP lambda); + DP XXZ_da1dlambda_kernel (DP sinzeta, DP coszeta, DP lambda); + DP XXZ_a2_kernel (DP sin2zeta, DP cos2zeta, DP lambda); + Root_Density XXZ_rhotot_GS (DP Delta, DP B, DP lambdamax, int Npts, DP req_prec); + DP Return_GS_Sz_tot_value (DP B, Root_Density& rhotot_GS); + Root_Density XXZ_eps_GS (DP Delta, DP Hz, DP lambdamax, int Npts, DP req_prec); + Root_Density XXZ_depsdlambda_GS (DP Delta, DP B, DP lambdamax, int Npts, DP req_prec); + Root_Density XXZ_b2BB_lambda_B (DP Delta, DP B, DP lambdamax, int Npts, DP req_prec); + Root_Density XXZ_b2BB_lambda_lambdap (DP Delta, DP B, DP lambdap, DP lambdamax, int Npts, DP req_prec); + Root_Density XXZ_Kbackflow_GS (DP Delta, DP B, DP lambdamax, DP lambda_p, DP lambda_h, int Npts, DP req_prec); + Root_Density XXZ_Fbackflow_GS (DP Delta, DP B, DP lambdamax, DP lambda_p, DP lambda_h, int Npts, DP req_prec); + Root_Density XXZ_Z_GS (DP Delta, DP B, DP lambdamax, int Npts, DP req_prec); + //void XXZ_Compare_Lattice_and_Continuum_Backflows_base_1010 (DP Delta, int N, int M, long long int id); + + // Defined in TBA_2CBG.cc: + struct TBA_Data_2CBG { + DP c_int; + DP mu; + DP Omega; + DP kBT; + DP f; // Gibbs free energy + DP n1; // first population + DP n2; // second population + }; + + // Defined in src/TBA_2CBG.cc: + TBA_Data_2CBG Solve_2CBG_TBAE_via_refinements (DP c_int, DP mu, DP Omega, DP kBT, int Max_Secs, bool Save_data); + + // Defined in src/TBA_2CBG.cc: + void Scan_2CBG_TBAE (DP c_int, DP mu_min, DP mu_max, int Nmu, DP Omega_min, DP Omega_max, int NOmega, + DP kBT_min, DP kBT_max, int NkBT, int Max_Secs); + + +} // namespace JSC + +#endif diff --git a/include/JSC_Vect.h b/include/JSC_Vect.h new file mode 100644 index 0000000..10feb46 --- /dev/null +++ b/include/JSC_Vect.h @@ -0,0 +1,472 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: JSC_Vect.h + +Purpose: Declares vector class. + +***********************************************************/ + +#ifndef _JSC_VECT_ +#define _JSC_VECT_ + +namespace JSC { + + template + class Vect { + private: + int dim; + T* V; + public: + Vect(); + explicit Vect (int N); + Vect (const T& a, int N); // initialize all N elements are a + Vect (const T* a, int N); // initialize to array + Vect (const Vect& rhs); // Copy constructor + Vect& operator= (const Vect& rhs); // assignment + Vect& operator= (const T& a); // assign a to all elements + inline T& operator[] (const int i); + inline const T& operator[] (const int i) const; + Vect& operator+= (const Vect& rhs); + Vect& operator-= (const Vect& rhs); + bool operator== (const Vect& rhs); // checks equality of size and of all elements + bool operator!= (const Vect& rhs); // checks inequality + bool Append (const T& rhs); // appends rhs to the vector + bool Append (const Vect& rhs); // appends rhs to the vector + bool Increase_Size (int nr_to_add); // resizes the array to accommodate nr_to_add more entries + bool Increase_Size (int nr_to_add, T setval); // resizes the array to accommodate nr_to_add more entries + inline int size() const; + inline double norm() const; // returns norm of vector + inline T max() const; // returns maximal value + inline T min() const; // returns maximal value + inline T sum() const; // returns sum of all elements + inline bool includes(T check) const; // whether check == one of the elements or not + void QuickSort (int l, int r); + void QuickSort (Vect& index, int l, int r); + void QuickSort (); + void QuickSort (Vect& index); + ~Vect(); + }; + + template + Vect::Vect() : dim(0), V(0) {} + + template + Vect::Vect (int N) : dim(N), V(new T[N]) {} + + template + Vect::Vect (const T& a, int N) : dim(N), V(new T[N]) + { + for (int i = 0; i < N; ++i) V[i] = a; + } + + template + Vect::Vect (const T* a, int N) : dim(N), V(new T[N]) + { + for (int i = 0; i < N; ++i) V[i] = *a++; + } + + template + Vect::Vect (const Vect& rhs) : dim(rhs.dim), V(new T[dim]) + { + for (int i = 0; i < dim; ++i) V[i] = rhs[i]; + } + + template + Vect& Vect::operator= (const Vect& rhs) + { + if (this != &rhs) { + if (dim != rhs.dim) { + if (V != 0) delete[] V; + dim = rhs.dim; + V = new T[dim]; + } + for (int i = 0; i < dim; ++i) V[i] = rhs[i]; + } + return *this; + } + + template + Vect& Vect::operator= (const T& a) + { + for (int i = 0; i < dim; ++i) V[i] = a; + return *this; + } + + template + inline T& Vect::operator[] (const int i) + { + return V[i]; + } + + template + inline const T& Vect::operator[] (const int i) const + { + return V[i]; + } + + template + Vect& Vect::operator+= (const Vect& rhs) + { + for (int i = 0; i < dim; ++i) V[i] += rhs[i]; + return *this; + } + + template + Vect& Vect::operator-= (const Vect& rhs) + { + for (int i = 0; i < dim; ++i) V[i] -= rhs[i]; + return *this; + } + + template + bool Vect::operator== (const Vect& rhs) + { + bool answer = ((*this).size() == rhs.size()); + if (answer) { + for (int i = 0; i < dim; ++i) answer = (answer && (V[i] == rhs[i])); + } + return answer; + } + + template + bool Vect::operator!= (const Vect& rhs) + { + return(!((*this) == rhs)); + } + + template + bool Vect::Append (const Vect& rhs) // appends rhs to the vector + { + T* newvect = new T[dim + rhs.size()]; + for (int i = 0; i < dim; ++i) newvect[i] = V[i]; + for (int i = 0; i < rhs.size(); ++i) newvect[i+ dim] = rhs[i]; + + dim += rhs.size(); + delete[] V; + V = new T[dim]; + for (int i = 0; i < dim; ++i) V[i] = newvect[i]; + + delete[] newvect; + + return(true); + } + + template + bool Vect::Append (const T& rhs) // appends rhs to the vector + { + T* newvect = new T[dim + 1]; + for (int i = 0; i < dim; ++i) newvect[i] = V[i]; + newvect[dim] = rhs; + + dim += 1; + delete[] V; + V = new T[dim]; + for (int i = 0; i < dim; ++i) V[i] = newvect[i]; + + delete[] newvect; + + return(true); + } + + template + bool Vect::Increase_Size (int nr_to_add) // resizes the array to accommodate nr_to_add more entries + { + int resized_dim = dim + nr_to_add; + + T* resized_vect = new T[resized_dim]; + for (int i = 0; i < dim; ++i) resized_vect[i] = V[i]; + for (int i = dim; i < resized_dim; ++i) resized_vect[i] = T(0); + + dim = resized_dim; + delete[] V; + V = new T[dim]; + for (int i = 0; i < dim; ++i) V[i] = resized_vect[i]; + + delete[] resized_vect; + + return(true); + } + + template + bool Vect::Increase_Size (int nr_to_add, T setval) // resizes the array to accommodate nr_to_add more entries + { + int resized_dim = dim + nr_to_add; + + T* resized_vect = new T[resized_dim]; + for (int i = 0; i < dim; ++i) resized_vect[i] = V[i]; + for (int i = dim; i < resized_dim; ++i) resized_vect[i] = setval; + + dim = resized_dim; + delete[] V; + V = new T[dim]; + for (int i = 0; i < dim; ++i) V[i] = resized_vect[i]; + + delete[] resized_vect; + + return(true); + } + + + template + inline int Vect::size() const + { + return dim; + } + + template + inline double Vect::norm () const + { + double normsq = 0.0; + for (int i = 0; i < dim; ++i) normsq += abs(V[i]) * abs(V[i]); + return sqrt(normsq); + } + + template <> + inline double Vect::norm () const + { + double normsq = 0.0; + for (int i = 0; i < dim; ++i) normsq += V[i] * V[i]; + return(sqrt(normsq)); + } + + template <> + inline double Vect >::norm () const + { + double normsq = 0.0; + for (int i = 0; i < dim; ++i) normsq += std::norm(V[i]); + return(sqrt(normsq)); + } + + template + inline T Vect::max() const + { + T maxval = V[0]; + for (int i = 0; i < dim; ++i) if (V[i] > maxval) maxval = V[i]; + return maxval; + } + + template + inline T Vect::min() const + { + T minval = V[0]; + for (int i = 0; i < dim; ++i) if (V[i] < minval) minval = V[i]; + return minval; + } + + template + inline T Vect::sum() const + { + T total = T(0); + for (int i = 0; i < dim; ++i) total += V[i]; + return total; + } + + template + inline bool Vect::includes (T check) const + { + int index = 0; + while (index < dim && V[index] != check) index++; + + return(index < dim); + } + /* + template + void Vect::QuickSort (int l, int r) + { + //cout << "QuickSort called for l = " << l << "\t r = " << r << endl; + //cout << (*this) << endl; + //for (int ih = l; ih <= r; ++ih) cout << setprecision(16) << "ih = " << ih << "\tV[ih] = " << V[ih] << endl; + + static T m; + static int j; + int i; + + if (r > l) { + m = V[r]; i = l-1; j = r; + + for (;;) { + while (V[++i] < m); + while (V[--j] > m); + if (i >= j) break; + std::swap(V[i], V[j]); + } + std::swap(V[i], V[r]); + + (*this).QuickSort(l, i-1); + (*this).QuickSort(i+1, r); + } + } + */ + /* + template + void Vect::QuickSort (int l, int r) + { + // My own version of QuickSort: add sentinels on left and right + if (r > l) { + int s = l + (r-l)/2; // central element index + // Rearrange so that V[l] <= V[s] <= V[r] (sentinels on left and right) + if (V[l] > V[r]) std::swap(V[l],V[r]); + if (V[s] > V[r]) std::swap(V[s],V[r]); + if (V[l] > V[s]) std::swap(V[l],V[s]); + m = V[s]; i = l-1; j = r; + //m = V[r]; i = l-1; j = r; + + for (;;) { + while (V[i] < m) i++; + while (V[j] > m) j--; + if (i >= j) break; + std::swap(V[i], V[j]); // restart from indices i and j just used, in case one is pivot + } + //std::swap(V[i], V[r]); + + (*this).QuickSort(l, i-1); + (*this).QuickSort(i+1, r); + } + + } + */ + + template + void Vect::QuickSort (int l, int r) + { + int i = l, j = r; + T pivot = V[l + (r-l)/2]; + + while (i <= j) { + while (V[i] < pivot) i++; + while (V[j] > pivot) j--; + if (i <= j) { + std::swap(V[i],V[j]); + i++; + j--; + } + }; + + if (l < j) (*this).QuickSort(l, j); + if (i < r) (*this).QuickSort(i, r); + } + + template + void Vect::QuickSort () + { + if ((*this).size() > 1) (*this).QuickSort (0, (*this).size() - 1); + } + + /* + template + void Vect::QuickSort (Vect& index, int l, int r) + { + if (index.size() != (*this).size()) { + cout << (*this).size() << "\t" << index.size() << endl; + JSCerror("Wrong dim for index in Vect QuickSort."); + } + + static T m; + static int j; + int i; + + if (r > l) { + m = V[r]; i = l-1; j = r; + + for (;;) { + while (V[++i] < m); + while (V[--j] > m); + if (i >= j) break; + std::swap(V[i], V[j]); + std::swap(index[i], index[j]); + } + std::swap(V[i], V[r]); + std::swap(index[i], index[r]); + + (*this).QuickSort(index, l, i-1); + (*this).QuickSort(index, i+1, r); + } + } + */ + /* + template + void Vect::QuickSort (Vect& index, int l, int r) + { + // My own version of QuickSort: + if (r > l) { + int s = l + (r-l)/2; // central element index + // Rearrange so that V[l] <= V[s] <= V[r] (sentinels on left and right) + if (V[l] > V[r]) std::swap(V[l],V[r]); + if (V[s] > V[r]) std::swap(V[s],V[r]); + if (V[l] > V[s]) std::swap(V[l],V[s]); + m = V[s]; i = l-1; j = r+1; + + for (;;) { + while (V[++i] < m); + while (V[--j] > m); + if (i >= j) break; + std::swap(index[i], index[j]); + std::swap(V[i--], V[j++]); // restart from indices i and j just used, in case one is pivot + } + + (*this).QuickSort(index, l, i-1); + (*this).QuickSort(index, i+1, r); + } + + } + */ + + template + void Vect::QuickSort (Vect& index, int l, int r) + { + int i = l, j = r; + T pivot = V[l + (r-l)/2]; + + while (i <= j) { + while (V[i] < pivot) i++; + while (V[j] > pivot) j--; + if (i <= j) { + std::swap(V[i],V[j]); + std::swap(index[i],index[j]); + i++; + j--; + } + }; + + if (l < j) (*this).QuickSort(index, l, j); + if (i < r) (*this).QuickSort(index, i, r); + } + + template + void Vect::QuickSort (Vect& index) + { + if (index.size() != (*this).size()) JSCerror("Wrong dim for index in Vect QuickSort."); + (*this).QuickSort (index, 0, (*this).size() - 1); + } + + template + inline std::ostream& operator<< (std::ostream& s, const Vect& vector) + { + for (int i = 0; i < vector.size() - 1; ++i) s << vector[i] << " "; + if (vector.size() >= 1) s << vector[vector.size() - 1]; + + return (s); + } + + template + Vect::~Vect() + { + if (V != 0) delete[] V; + } + + + // TYPEDEFS + typedef JSC::Vect Vect_INT; + typedef JSC::Vect Vect_DP; + typedef JSC::Vect > Vect_CX; + + +} // namespace JSC + +#endif diff --git a/include/JSC_XXX_h0.h b/include/JSC_XXX_h0.h new file mode 100644 index 0000000..020dd88 --- /dev/null +++ b/include/JSC_XXX_h0.h @@ -0,0 +1,139 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_XXX_h0.h + +Purpose: Declares classes for XXX in zero field: Uq(sl(2)) stuff. + + +***********************************************************/ + +#ifndef _XXX_h0_ +#define _XXX_h0_ + +#include "JSC.h" + +const DP default_req_prec = 1.0e-14; +const int default_max_rec = 10; + +namespace JSC { + + inline DP Integrand_11 (Vect_DP args) + { + // args[0] corresponds to t, args[1] to rho + return((exp(args[0]) * sinh(args[0]) * cos(4.0 * args[0] * args[1])/(2.0 * pow(cosh(args[0]), 2.0)) + 2.0 * pow(sin(2.0 * args[0] * args[1]), 2.0))/args[0]); + } + + inline DP Integrand_12 (Vect_DP args) + { + DP expm2t = exp(-2.0*args[0]); + return(cos(4.0 * args[0] * args[1]) * expm2t * (3.0 + expm2t)/ (args[0] * (1.0 + expm2t) * (1.0 + expm2t))); + } + + inline DP Integrand_2 (Vect_DP args) + { + DP answer = 0.0; + if (args[0] < 1.0) answer = exp(args[0]) * pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * sinh(args[0]) * pow(cosh(args[0]), 2.0)); + else if (args[0] >= 1.0) { + DP expm2t = exp(-2.0 * args[0]); + answer = 8.0 * expm2t * pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * (1.0 - expm2t) * (1.0 + expm2t) * (1.0 + expm2t)); + } + return(answer); + } + + inline DP Integrand_A (Vect_DP args) + { + // This kernel is for -ln | A_-(i\pi/2) | function + return(exp(args[0]) * pow(sinh(args[0]/2.0), 2.0)/(args[0] * sinh(2.0*args[0]) * cosh(args[0]))); + } + + DP I_integral (DP rho, DP req_prec); + + + /********************* TWO SPINONS ********************/ + + DP SF_2p (DP k, DP omega, I_table Itable); + DP SF_2p (Vect_DP args, I_table Itable); + DP SF_2p_alt (Vect_DP args, I_table Itable); + DP SF_2p_w (Vect_DP args, I_table Itable); + DP SF_2p_w_alt (Vect_DP args, I_table Itable); + DP SF_2p_intw (Vect_DP args, I_table Itable); + DP SF_2p_intw_alt (Vect_DP args, I_table Itable); + DP SF_2p_check_sumrule (DP req_prec, int max_rec, I_table Itable); + DP SF_2p_check_sumrule_alt (DP req_prec, int max_rec, I_table Itable); + DP Fixed_k_sumrule_w (DP k); + DP Fixed_k_sumrule_omega (DP k); + DP SF_2p_check_fixed_k_sumrule (DP k, DP req_prec, int max_rec, I_table Itable); + DP SF_2p_check_fixed_k_sumrule_alt (DP k, DP req_prec, int max_rec, I_table Itable); + DP SF_2p_check_fixed_k_sumrule_opt (DP k, DP req_prec, int Npts, I_table Itable); + + +/********************** FOUR SPINONS **********************/ + + DP Sum_norm_gl (Vect_DP rho, DP req_prec); + DP Compute_C4 (DP req_prec); + DP SF_contrib (Vect_DP p, DP req_prec, I_table Itable); + DP J_fn (Vect_DP p, DP req_prec, I_table Itable); + inline DP Jacobian_p3p4_KW (DP k, DP w, DP K, DP W) + { + return(1.0/sqrt(pow(twoPI * sin(0.5 * (k - K)), 2.0) - (w-W)*(w-W))); + } + bool Set_p_given_kwKW (DP k, DP w, DP K, DP W, Vect_DP& p); + inline DP wmin_4p (DP k) + { + return(PI * fabs(sin(k))); + } + inline DP wmax_4p (DP k) + { + return(2.0 * PI * sqrt(2.0 * (1.0 + fabs(cos(0.5*k))))); + } + inline DP Wmin (DP k, DP w, DP K) + { + return(JSC::max(1.0e-15, JSC::max(fabs(PI * sin(K)), w - twoPI * sin(0.5 * (fabs(k-K)))))); + } + inline DP Wmax (DP k, DP w, DP K) + { + return(JSC::min(twoPI * sin(0.5 * K), w - fabs(PI * sin(k - K)))); + } + DP G_fn (Vect_DP args_to_G, I_table Itable); + DP G1_fn (Vect_DP args_to_G, I_table Itable); + DP G2_fn (Vect_DP args_to_G, I_table Itable); + DP G1_fn_mid (Vect_DP args_to_G, I_table Itable); + DP G2_fn_mid (Vect_DP args_to_G, I_table Itable); + DP G_fn_alt (Vect_DP args_to_G, I_table Itable); + DP H_fn (Vect_DP args_to_H, I_table Itable); + DP H2_fn (Vect_DP args_to_H, I_table Itable); + DP H_fn_mid (Vect_DP args_to_H, I_table Itable); + DP H_fn_alt (Vect_DP args_to_H, I_table Itable); + DP SF_4p_kwKW (Vect_DP args, I_table Itable); + DP SF_4p_kwKW_alpha (Vect_DP args, I_table Itable); + DP SF_4p_kwKW_alpha_opt (Vect_DP args, I_table Itable); + + // Interface to used version: + DP SF_4p_rec (DP k, DP omega, DP req_prec, int max_rec, I_table Itable); + DP SF_4p (DP k, DP omega, I_table Itable); + DP SF_4p_opt (DP k, DP omega, DP req_prec, int Npts_K, int Npts_W, I_table Itable); + void Translate_raw_4p_data (DP k, int max_rec_w, const char* SFraw_Cstr, const char* SF_Cstr, const char* SFsrc_Cstr, I_table Itable); + DP SF_4p_rec (DP k, DP req_prec, int max_rec_w, int max_rec, I_table Itable); + Integral_result SF_4p_opt (DP k, DP req_prec, int Npts, I_table Itable); + Integral_result SF_4p_opt (DP k, DP req_prec, int Npts_w, int Npts_KW, I_table Itable); + Integral_result SF_4p_opt (DP k, DP req_prec, int Npts_w, int Npts_K, int Npts_W, I_table Itable); + + //******************************** Functions to produce files similar to ABACUS ********************************** + void Produce_SF_2p_file (int N, int Nomega, DP omegamax, I_table Itable); + void Produce_SF_4p_file (int N, int Nomega, DP omegamax, DP req_prec, int max_rec, I_table Itable); + void Produce_SF_4p_file (int N, int Nomega, DP omegamax, DP req_prec, int Npts_K, int Npts_W, I_table Itable); + + //******************************** New version, full k and omega integral in one go ****************************** + DP Direct_J_integral (int Npts_p, DP req_prec, I_table Itable); + DP Direct_J_integral_bin (int Npts_p, int Npts_o, DP req_prec, I_table Itable); + void Smoothen_raw_SF_4p (int Npts_p, int Npts_o, DP req_prec, DP width); + +} // namespace JSC + +#endif diff --git a/include/JSC_XXZ_h0.h b/include/JSC_XXZ_h0.h new file mode 100644 index 0000000..cc5e378 --- /dev/null +++ b/include/JSC_XXZ_h0.h @@ -0,0 +1,86 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_XXZ_h0.h + +Purpose: Declares classes for XXZ in zero field: quantum group stuff. + + +***********************************************************/ + +#ifndef _XXZ_h0_ +#define _XXZ_h0_ + +#include "JSC.h" + +//const DP default_req_prec = 1.0e-14; +//const int default_max_rec = 10; + +namespace JSC { + + /* + inline DP Integrand_xi_11 (Vect_DP args) + { + // args[0] corresponds to t, args[1] to rho, args[2] to xi + //return((exp(args[0]) * sinh(args[0]) * cos(4.0 * args[0] * args[1])/(2.0 * pow(cosh(args[0]), 2.0)) + 2.0 * pow(sin(2.0 * args[0] * args[1]), 2.0))/args[0]); + return((sinh(args[0]*(1.0 + args[2])) * sinh(args[0]) * cos(4.0 * args[0] * args[1])/(2.0 * sinh(args[0] * args[2]) * pow(cosh(args[0]), 2.0)) + + 2.0 * pow(sin(2.0 * args[0] * args[1]), 2.0))/args[0]); + } + + inline DP Integrand_xi_12 (Vect_DP args) + { + DP expm2t = exp(-2.0*args[0]); + DP expm2txi = exp(-2.0*args[0]*args[2]); + //return(cos(4.0 * args[0] * args[1]) * expm2t * (3.0 + expm2t)/ (args[0] * (1.0 + expm2t) * (1.0 + expm2t))); + return(cos(4.0 * args[0] * args[1]) * (expm2t * (3.0 * (1.0 + expm2txi) + expm2t) + expm2txi) / (args[0] * (1.0 - expm2txi) * (1.0 + expm2t) * (1.0 + expm2t))); + } + */ + /* + inline DP Integrand_xi_2 (Vect_DP args) + { + DP answer = 0.0; + if (args[0] < 1.0) + //answer = exp(args[0]) * pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * sinh(args[0]) * pow(cosh(args[0]), 2.0)); + answer = sinh(args[0] * (args[2] + 1.0)) * pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * sinh(args[2] * args[0]) * sinh(args[0]) * pow(cosh(args[0]), 2.0)); + else if (args[0] >= 1.0) { + DP expm2t = exp(-2.0 * args[0]); + DP expm2txi = exp(-2.0*args[0]*args[2]); + //answer = 8.0 * expm2t * pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * (1.0 - expm2t) * (1.0 + expm2t) * (1.0 + expm2t)); + answer = 8.0 * ((1.0 - expm2t*expm2txi)/(1.0 - expm2t*expm2txi)) * expm2t * + pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * (1.0 - expm2t) * (1.0 + expm2t) * (1.0 + expm2t)); + } + return(answer); + } + */ + + DP I_xi_integral (DP xi, DP rho, DP req_prec, int max_nr_pts); + + + /********************* TWO SPINONS ********************/ + DP Szz_XXZ_h0_2spinons (DP k, DP omega, Integral_table Itable); + DP Szz_XXZ_h0_2spinons (Vect_DP args, Integral_table Itable); + DP Szz_XXZ_h0_2spinons_alt (Vect_DP args, Integral_table Itable); + DP Szz_XXZ_h0_2spinons_omega (Vect_DP args, Integral_table Itable); + DP Szz_XXZ_h0_2spinons_omega_alt (Vect_DP args, Integral_table Itable); + DP Szz_XXZ_h0_2spinons_intomega (Vect_DP args, Integral_table Itable); + DP Szz_XXZ_h0_2spinons_intomega_alt (Vect_DP args, Integral_table Itable); + DP Szz_XXZ_h0_2spinons_check_sumrule (DP Delta, DP req_prec, int max_nr_pts, Integral_table Itable); + DP Szz_XXZ_h0_2spinons_check_sumrule_alt (DP Delta, DP req_prec, int max_nr_pts, Integral_table Itable); + DP Fixed_k_sumrule_omega_Szz_XXZ_h0_N (DP Delta, DP k); + DP GSE_XXZ_h0 (DP Delta, DP req_prec, int max_nr_pts); + DP Fixed_k_sumrule_omega_Szz_XXZ_h0 (DP Delta, DP k, DP req_prec, int max_nr_pts); + DP Szz_XXZ_h0_2spinons_check_fixed_k_Szz_sumrule (DP Delta, DP k, DP req_prec, int max_nr_pts, Integral_table Itable); + DP Szz_XXZ_h0_2spinons_check_fixed_k_Szz_sumrule_alt (DP Delta, DP k, DP req_prec, int max_nr_pts, Integral_table Itable); + + //******************************** Functions to produce files similar to ABACUS ********************************** + void Produce_Szz_XXZ_h0_2spinons_file (DP Delta, int N, int Nomega, DP omegamax, Integral_table Itable); + void Produce_Szz_XXZ_h0_2spinons_fixed_K_file (DP Delta, DP Kover2PI, int Nomega, Integral_table Itable); + +} // namespace JSC + +#endif diff --git a/include/JSC_Young.h b/include/JSC_Young.h new file mode 100644 index 0000000..e50abf8 --- /dev/null +++ b/include/JSC_Young.h @@ -0,0 +1,118 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_Young.h + +Purpose: Declares Young tableau class. + +***********************************************************/ + +#ifndef _YOUNG_ +#define _YOUNG_ + +#include "JSC_Vect.h" + +namespace JSC { + + const int YOUNG_TABLEAU_ID_OPTION = 0; + const long long int TABLEAU_ID_UPPER_LIMIT = 10000000LL; + + //*********************************************************************** + + class Young_Tableau { + + public: + int Nrows; + int Ncols; + int* Row_L; + int* Col_L; + long long int id; // identification number + long long int maxid; + long long int* map; + bool map_computed; + long long int idnr_reached; + int nboxes_reached; + + private: + int dimchoose; + long long int* choose_table; + + public: + Young_Tableau (); // empty constructor, does nothing + Young_Tableau (int Nr, int Nc); // constructs empty tableau + Young_Tableau (int Nr, int Nc, long long int idnr); // constructs the tableau corresponding to identification number idnr + Young_Tableau (const Young_Tableau& RefTableau); // copy constructor + Young_Tableau (int Nr, int Nc, const Young_Tableau& RefTableau); + Young_Tableau& operator= (const Young_Tableau& RefTableau); // assignment + ~Young_Tableau (); // destructor + + public: + Young_Tableau& Compute_Map (long long int idnr_to_reach); // fills the map vector + Young_Tableau& Distribute_boxes (int nboxes_to_dist, int level); + Young_Tableau& Set_to_id (long long int idnr); // sets the tableau to the one corresponding to idnr + Young_Tableau& Set_to_id (long long int idnr, int option); // sets the tableau to the one corresponding to idnr according to rule option + Young_Tableau& Set_Row_L (Vect& Row_Lengths); // set row lengths + Young_Tableau& Set_Col_L_given_Row_L (); // sets the Col_L array self-consistently + Young_Tableau& Set_Row_L_given_Col_L (); // sets the Col_L array self-consistently + long long int Compute_Descendent_id (int option, Vect& Desc_Row_L, int Nrows_Desc, int Ncols_Desc, + const Young_Tableau& RefTableau); + Young_Tableau& Compute_id(); // computes the id number of tableau + Young_Tableau& Compute_id(int option); // computes the id number of tableau according to rule option + Young_Tableau& Print(); // couts the tableau + + bool Lower_Row (int i); + bool Raise_Row (int i); + bool Lower_Col (int i); + bool Raise_Col (int i); + bool Raise_Lowest_Nonzero_Row(); // adds a box to the lowest nonzero length Row, recomputes id, returns true if tableau has changed + bool Raise_Next_to_Lowest_Nonzero_Row(); // same thing, but for Row under lowest nonzero length one. + bool Move_Box_from_Col_to_Col (int ifrom, int ito); + + Vect Descendents (int fixed_Nboxes); + Vect Descendents_Boosted_State (int fixed_Nboxes); + + int Add_Boxes_From_Lowest (int Nboxes); // tries to add Nboxes to Tableau, returns number of boxes added. + }; + + std::ostream& operator<< (std::ostream& s, const Young_Tableau& tableau); + + inline int Nr_Nonzero_Rows (const Vect& Tableau_ref) + { + // This function checks the number of rows containing at least one box + // in the whole vector of Young tableaux given as argument. + // The usefulness is to force descent of states in which only a few + // excitations have started dispersing. + + int nr_nonzero_rows = 0; + for (int i = 0; i < Tableau_ref.size(); ++i) + for (int alpha = 0; alpha < Tableau_ref[i].Nrows; ++alpha) + if (Tableau_ref[i].Row_L[alpha] > 0) nr_nonzero_rows++; + + return(nr_nonzero_rows); + } + + //*********************************************************************** + + class Tableau_Map { + + public: + Vect map; + long long int idnr_reached; + int nboxes_reached; + + public: + Tableau_Map (int Nrows, int Ncols); + void Distribute_id (int nboxes_to_dist, int level, Young_Tableau& RefTableau); + + }; + + + +} // namespace JSC + +#endif diff --git a/include/JSC_util.h b/include/JSC_util.h new file mode 100644 index 0000000..b3f5f44 --- /dev/null +++ b/include/JSC_util.h @@ -0,0 +1,485 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: JSC_util.h + +Purpose: Defines basic math functions. + +***********************************************************/ + +#ifndef _JSC_UTIL_H_ +#define _JSC_UTIL_H_ + +#include "JSC.h" + +using namespace std; + +typedef double DP; + +// Global constants + +const double PI = 3.141592653589793238462643; +const double sqrtPI = sqrt(PI); +const double twoPI = 2.0*PI; +const double logtwoPI = log(twoPI); +const double Euler_Mascheroni = 0.577215664901532860606; +const double Gamma_min_0p5 = -2.0 * sqrt(PI); +const complex II(0.0,1.0); // Shorthand for i + +const DP MACHINE_EPS = numeric_limits::epsilon(); +const DP MACHINE_EPS_SQ = pow(MACHINE_EPS, 2.0); + +// Now for some basic math utilities: + +namespace JSC { + + // File checks: + + inline bool file_exists (const char* filename) + { + fstream file; + file.open(filename); + bool exists = !file.fail(); + file.close(); + + return(exists); + } + + // Error handler: + + inline void JSCerror (const string error_text) + // my error handler + { + cerr << "Run-time error... " << endl; + cerr << error_text << endl; + cerr << "Exiting to system..." << endl; + exit(1); + } + + struct Divide_by_zero {}; + + + // Basics: min, max, fabs + + template + inline const T max (const T& a, const T& b) { return a > b ? (a) : (b); } + + template + inline const T min (const T& a, const T& b) { return a > b ? (b) : (a); } + + template + inline const T fabs (const T& a) { return a >= 0 ? (a) : (-a); } + + inline long long int pow_lli (const long long int& base, const int& exp) + { + long long int answer = base; + if (exp == 0) answer = 1LL; + else for (int i = 1; i < exp; ++i) answer *= base; + return(answer); + } + + inline unsigned long long int pow_ulli (const unsigned long long int& base, const int& exp) + { + unsigned long long int answer = base; + if (exp == 0) answer = 1ULL; + for (int i = 1; i < exp; ++i) answer *= base; + return(answer); + } + + inline int fact (const int& N) + { + int ans = 0; + + if (N < 0) { + cerr << "Error: factorial of negative number. Exited." << endl; + exit(1); + } + else if ( N == 1 || N == 0) ans = 1; + else ans = N * fact(N-1); + + return(ans); + } + + inline DP ln_fact (const int& N) + { + DP ans = 0.0; + + if (N < 0) { + cerr << "Error: factorial of negative number. Exited." << endl; + exit(1); + } + else if ( N == 1 || N == 0) ans = 0.0; + else ans = log(DP(N)) + ln_fact(N-1); + + return(ans); + } + + inline long long int fact_lli (const int& N) + { + long long int ans = 0; + + if (N < 0) { + cerr << "Error: factorial of negative number. Exited." << endl; + exit(1); + } + else if ( N == 1 || N == 0) ans = 1; + else ans = fact_lli(N-1) * N; + + return(ans); + } + + inline long long int fact_ulli (const int& N) + { + unsigned long long int ans = 0; + + if (N < 0) { + cerr << "Error: factorial of negative number. Exited." << endl; + exit(1); + } + else if ( N == 1 || N == 0) ans = 1; + else ans = fact_ulli(N-1) * N; + + return(ans); + } + + inline int choose (const int& N1, const int& N2) + { + // returns N1 choose N2 + + int ans = 0; + if (N1 < N2) { + cout << "Error: N1 smaller than N2 in choose. Exited." << endl; + exit(1); + } + else if (N1 == N2) ans = 1; + else if (N1 < 12) ans = fact(N1)/(fact(N2) * fact(N1 - N2)); + else { + ans = 1; + int mult = N1; + while (mult > max(N2, N1 - N2)) ans *= mult--; + ans /= fact(min(N2, N1 - N2)); + } + + return(ans); + } + + inline DP ln_choose (const int& N1, const int& N2) + { + // returns the log of N1 choose N2 + + DP ans = 0.0; + if (N1 < N2) { + cout << "Error: N1 smaller than N2 in choose. Exited." << endl; + exit(1); + } + else if (N1 == N2) ans = 0.0; + else ans = ln_fact(N1) - ln_fact(N2) - ln_fact(N1 - N2); + + return(ans); + } + + + inline long long int choose_lli (const int& N1, const int& N2) + { + // returns N1 choose N2 + + long long int ans = 0; + if (N1 < N2) { + cout << "Error: N1 smaller than N2 in choose. Exited." << endl; + exit(1); + } + else if (N1 == N2) ans = 1; + else if (N1 < 12) ans = fact_lli(N1)/(fact_lli(N2) * fact_lli(N1 - N2)); + else { + // Make sure that N2 is less than or equal to N1/2; if not, just switch... + int N2_min = min(N2, N1 - N2); + + ans = 1; + for (int i = 0; i < N2_min; ++i) { + ans *= (N1 - i); + ans /= i + 1; + } + } + + return(ans); + } + + inline unsigned long long int choose_ulli (const int& N1, const int& N2) + { + // returns N1 choose N2 + + unsigned long long int ans = 0; + if (N1 < N2) { + cout << "Error: N1 smaller than N2 in choose. Exited." << endl; + exit(1); + } + else if (N1 == N2) ans = 1; + else if (N1 < 12) ans = fact_ulli(N1)/(fact_ulli(N2) * fact_ulli(N1 - N2)); + else { + // Make sure that N2 is less than or equal to N1/2; if not, just switch... + int N2_min = min(N2, N1 - N2); + + ans = 1; + for (int i = 0; i < N2_min; ++i) { + ans *= (N1 - i); + ans /= i + 1; + } + } + + return(ans); + } + + inline DP SIGN (const DP &a, const DP &b) + { + return b >= 0 ? (a >= 0 ? a : -a) : (a >= 0 ? -a : a); + } + + inline DP sign_of (const DP& a) + { + return (a >= 0.0 ? 1.0 : -1.0); + } + + inline int sgn_int (const int& a) + { + return (a >= 0) ? 1 : -1; + } + + inline int sgn_DP (const DP& a) + { + return (a >= 0) ? 1 : -1; + } + + template + inline void SWAP (T& a, T& b) {T dum = a; a = b; b = dum;} + + inline int kronecker (int a, int b) + { + return a == b ? 1 : 0; + } + + template + inline bool is_nan (const T& a) + { + return(!((a < T(0.0)) || (a >= T(0.0)))); + } + + inline complex atan_cx(const complex& x) + { + return(-0.5 * II * log((1.0 + II* x)/(1.0 - II* x))); + } + + /**************** Gamma function *******************/ + + inline complex ln_Gamma (complex z) + { + // Implementation of Lanczos method with g = 9. + // Coefficients from Godfrey 2001. + + if (real(z) < 0.5) return(log(PI/(sin(PI*z))) - ln_Gamma(1.0 - z)); + + else { + + complex series = 1.000000000000000174663 + + 5716.400188274341379136/z + - 14815.30426768413909044/(z + 1.0) + + 14291.49277657478554025/(z + 2.0) + - 6348.160217641458813289/(z + 3.0) + + 1301.608286058321874105/(z + 4.0) + - 108.1767053514369634679/(z + 5.0) + + 2.605696505611755827729/(z + 6.0) + - 0.7423452510201416151527e-2 / (z + 7.0) + + 0.5384136432509564062961e-7 / (z + 8.0) + - 0.4023533141268236372067e-8 / (z + 9.0); + + return(0.5 * logtwoPI + (z - 0.5) * log(z + 8.5) - (z + 8.5) + log(series)); + } + + return(log(0.0)); // never called + } + + inline complex ln_Gamma_old (complex z) + { + // Implementation of Lanczos method with g = 9. + // Coefficients from Godfrey 2001. + + if (real(z) < 0.5) return(log(PI/(sin(PI*z))) - ln_Gamma(1.0 - z)); + + else { + + int g = 9; + + double p[11] = { 1.000000000000000174663, + 5716.400188274341379136, + -14815.30426768413909044, + 14291.49277657478554025, + -6348.160217641458813289, + 1301.608286058321874105, + -108.1767053514369634679, + 2.605696505611755827729, + -0.7423452510201416151527e-2, + 0.5384136432509564062961e-7, + -0.4023533141268236372067e-8 }; + + complex z_min_1 = z - 1.0; + complex series = p[0]; + for (int i = 1; i < g+2; ++i) + series += p[i]/(z_min_1 + complex(i)); + + return(0.5 * logtwoPI + (z_min_1 + 0.5) * log(z_min_1 + complex(g) + 0.5) - (z_min_1 + complex(g) + 0.5) + log(series)); + } + + return(log(0.0)); // never called + } + + inline complex ln_Gamma_2 (complex z) + { + // Implementation of Lanczos method with g = 7. + + if (real(z) < 0.5) return(log(PI/(sin(PI*z)) - ln_Gamma(1.0 - z))); + + else { + + int g = 7; + + double p[9] = { 0.99999999999980993, 676.5203681218851, -1259.1392167224028, + 771.32342877765313, -176.61502916214059, 12.507343278686905, + -0.13857109526572012, 9.9843695780195716e-6, 1.5056327351493116e-7}; + + complex z_min_1 = z - 1.0; + complex series = p[0]; + for (int i = 1; i < g+2; ++i) + series += p[i]/(z_min_1 + complex(i)); + + return(0.5 * logtwoPI + (z_min_1 + 0.5) * log(z_min_1 + complex(g) + 0.5) - (z_min_1 + complex(g) + 0.5) + log(series)); + } + + return(log(0.0)); // never called + } + + /********** Partition numbers **********/ + + inline long long int Partition_Function (int n) + { + // Returns the value of the partition function p(n), giving the number of partitions of n into integers. + + if (n < 0) JSCerror("Calling Partition_Function for n < 0."); + else if (n == 0 || n == 1) return(1LL); + else if (n == 2) return(2LL); + else if (n == 3) return(3LL); + + else { // do recursion using pentagonal numbers + long long int pn = 0LL; + int pentnrplus, pentnrmin; // pentagonal numbers + for (int i = 1; true; ++i) { + pentnrplus = (i * (3*i - 1))/2; + pentnrmin = (i * (3*i + 1))/2; + //cout << "\ti = " << i << "pnrp = " << pentnrplus << "\tpnrm = " << pentnrmin << endl; + if (n - pentnrplus >= 0) pn += (i % 2 ? 1LL : -1LL) * Partition_Function (n - pentnrplus); + if (n - pentnrmin >= 0) pn += (i % 2 ? 1LL : -1LL) * Partition_Function (n - pentnrmin); + else break; + } + return(pn); + } + return(-1LL); // never called + } + + + /********** Sorting **********/ + + /* + template + void QuickSort (item_type* a, int l, int r) + { + static item_type m; + static int j; + int i; + + if (r > l) { + m = a[r]; i = l-1; j = r; + for (;;) { + while (a[++i] < m); + while (a[--j] > m); + if (i >= j) break; + std::swap(a[i], a[j]); + } + std::swap(a[i], a[r]); + QuickSort(a, l, i-1); + QuickSort(a, i+1, r); + } + } + */ + + template + void QuickSort (T* V, int l, int r) + { + int i = l, j = r; + T pivot = V[l + (r-l)/2]; + + while (i <= j) { + while (V[i] < pivot) i++; + while (V[j] > pivot) j--; + if (i <= j) { + std::swap(V[i],V[j]); + i++; + j--; + } + }; + + if (l < j) QuickSort(V, l, j); + if (i < r) QuickSort(V, i, r); + } + + /* + template + void QuickSort (item_type* a, int* idx, int l, int r) + { + static item_type m; + static int j; + int i; + + if (r > l) { + m = a[r]; i = l-1; j = r; + for (;;) { + while (a[++i] < m); + while (a[--j] > m); + if (i >= j) break; + std::swap(a[i], a[j]); + std::swap(idx[i], idx[j]); + } + std::swap(a[i], a[r]); + std::swap(idx[i], idx[r]); + QuickSort(a, idx, l, i-1); + QuickSort(a, idx, i+1, r); + } + } + */ + template + void QuickSort (T* V, int* index, int l, int r) + { + int i = l, j = r; + T pivot = V[l + (r-l)/2]; + + while (i <= j) { + while (V[i] < pivot) i++; + while (V[j] > pivot) j--; + if (i <= j) { + std::swap(V[i],V[j]); + std::swap(index[i],index[j]); + i++; + j--; + } + }; + + if (l < j) QuickSort(V, index, l, j); + if (i < r) QuickSort(V, index, i, r); + } + + +} // namespace JSC + +#endif // _JS_UTIL_H_ diff --git a/src/BETHE/Base.cc b/src/BETHE/Base.cc new file mode 100644 index 0000000..db31676 --- /dev/null +++ b/src/BETHE/Base.cc @@ -0,0 +1,307 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +----------------------------------------------------------- + +File: src/BETHE/Base.cc + +Purpose: defines functions in Base class, + providing a unified base object for all + Bethe Ansatz integrable models. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + // Function definitions: class Base + + Base::Base () : Charge(0), Nrap(Vect()), Nraptot(0), Ix2_infty(Vect()), + Ix2_max(Vect()), id(0LL) {} + + Base::Base (int N) : Charge(N), Nrap(Vect(N,1)), Nraptot(N), Ix2_infty(Vect(1.0e+100,1)), + Ix2_max(Vect(LONG_LONG_MAX, 1)), id(N) {} + + Base::Base (const Base& RefBase) // copy constructor + : Charge(RefBase.Charge), Nrap(Vect(RefBase.Nrap.size())), Nraptot(RefBase.Nraptot), + Ix2_infty(Vect(RefBase.Ix2_infty.size())), Ix2_max(Vect(RefBase.Ix2_max.size())), id(RefBase.id) + { + for (int i = 0; i < Nrap.size(); ++i) { + Nrap[i] = RefBase.Nrap[i]; + Ix2_infty[i] = RefBase.Ix2_infty[i]; + Ix2_max[i] = RefBase.Ix2_max[i]; + } + } + + /* + // DEPRECATED + Base::Base (const Heis_Chain& RefChain, int M) + : Charge(M), Nrap(Vect(RefChain.Nstrings)), Nraptot(0), Ix2_infty(Vect(RefChain.Nstrings)), Ix2_max(Vect(RefChain.Nstrings)) + { + for (int i = 0; i < RefChain.Nstrings; ++i) Nrap[i] = 0; + Nrap[0] = M; + + Nraptot = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Nraptot += Nrap[i]; + + // The id of this is zero by definition + id = 0LL; + + // Now compute the Ix2_infty numbers + + (*this).Compute_Ix2_limits(RefChain); + + } + */ + Base::Base (const Heis_Chain& RefChain, const Vect& Nrapidities) + : Charge(0), Nrap(Nrapidities), Nraptot(0), Ix2_infty(Vect(RefChain.Nstrings)), Ix2_max(Vect(RefChain.Nstrings)), + id (0LL) + { + + // Check consistency of Nrapidities vector with RefChain + + //if (RefChain.Nstrings != Nrapidities.size()) cout << "error: Nstrings = " << RefChain.Nstrings << "\tNrap.size = " << Nrapidities.size() << endl; + + if (RefChain.Nstrings != Nrapidities.size()) JSCerror("Incompatible Nrapidities vector used in Base constructor."); + + int Mcheck = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Mcheck += RefChain.Str_L[i] * Nrap[i]; + Charge = Mcheck; + + Nraptot = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Nraptot += Nrap[i]; + + // Compute id + id += Nrapidities[0]; + long long int factor = 100000LL; + for (int i = 1; i < RefChain.Nstrings; ++i) { + id += factor * Nrapidities[i]; + factor *= 100LL; + } + + // Now compute the Ix2_infty numbers + + (*this).Compute_Ix2_limits(RefChain); + + } + + Base::Base (const Heis_Chain& RefChain, long long int id_ref) + : Charge(0), Nrap(Vect(RefChain.Nstrings)), Nraptot(0), Ix2_infty(Vect(RefChain.Nstrings)), Ix2_max(Vect(RefChain.Nstrings)), + id (id_ref) + { + // Build Nrapidities vector from id_ref + + long long int factor = pow_ulli (10LL, 2* RefChain.Nstrings + 1); + long long int id_eff = id_ref; + for (int i = 0; i < RefChain.Nstrings - 1; ++i) { + Nrap[RefChain.Nstrings - 1 - i] = id_eff/factor; + id_eff -= factor * Nrap[RefChain.Nstrings - 1 - i]; + factor /= 100LL; + } + Nrap[0] = id_eff; + + //id = id_ref; + + //cout << "In Base constructor: id_ref = " << id_ref << " and Nrapidities = " << Nrap << endl; + + // Check consistency of Nrapidities vector with RefChain + + //if (RefChain.Nstrings != Nrap.size()) JSCerror("Incompatible Nrapidities vector used in Base constructor."); + + int Mcheck = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Mcheck += RefChain.Str_L[i] * Nrap[i]; + Charge = Mcheck; + + Nraptot = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Nraptot += Nrap[i]; + + // Now compute the Ix2_infty numbers + + (*this).Compute_Ix2_limits(RefChain); + } + + + Base& Base::operator= (const Base& RefBase) + { + if (this != & RefBase) { + Charge = RefBase.Charge; + Nrap = RefBase.Nrap; + Nraptot = RefBase.Nraptot; + Ix2_infty = RefBase.Ix2_infty; + Ix2_max = RefBase.Ix2_max; + id = RefBase.id; + } + return(*this); + } + + bool Base::operator== (const Base& RefBase) + { + bool answer = (Nrap == RefBase.Nrap); + + return (answer); + } + + bool Base::operator!= (const Base& RefBase) + { + bool answer = (Nrap != RefBase.Nrap); + + return (answer); + } + + void Base::Compute_Ix2_limits (const Heis_Chain& RefChain) + { + + if ((RefChain.Delta > 0.0) && (RefChain.Delta < 1.0)) { + + // Compute the Ix2_infty numbers + + DP sum1 = 0.0; + DP sum2 = 0.0; + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + sum1 = 0.0; + + for (int k = 0; k < RefChain.Nstrings; ++k) { + + sum2 = 0.0; + + sum2 += (RefChain.Str_L[j] == RefChain.Str_L[k]) ? 0.0 : 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j] * RefChain.par[k]) + - 0.5 * fabs(RefChain.Str_L[j] - RefChain.Str_L[k]) * RefChain.anis)); + sum2 += 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j] * RefChain.par[k]) + - 0.5 * (RefChain.Str_L[j] + RefChain.Str_L[k]) * RefChain.anis)); + + for (int a = 1; a < JSC::min(RefChain.Str_L[j], RefChain.Str_L[k]); ++a) + sum2 += 2.0 * 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j] * RefChain.par[k]) + - 0.5 * (fabs(RefChain.Str_L[j] - RefChain.Str_L[k]) + 2.0*a) * RefChain.anis)); + + sum1 += (Nrap[k] - ((j == k) ? 1 : 0)) * sum2; + } + + Ix2_infty[j] = (1.0/PI) * fabs(RefChain.Nsites * 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j]) + - 0.5 * RefChain.Str_L[j] * RefChain.anis)) - sum1); + + } // The Ix2_infty are now set. + + // Now compute the Ix2_max limits + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + Ix2_max[j] = int(floor(Ix2_infty[j])); // sets basic integer + + // Reject formally infinite rapidities (i.e. if Delta is root of unity) + + //cout << "Ix2_infty - Ix2_max = " << Ix2_infty[j] - Ix2_max[j] << endl; + //if (Ix2_infty[j] == Ix2_max[j]) { + //Ix2_max[j] -= 2; + //} + // If Nrap is even, Ix2_max must be odd. If odd, then even. + + if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; + + while (Ix2_max[j] > RefChain.Nsites) { + Ix2_max[j] -= 2; + } + } + } // if XXZ gapless + + else if (RefChain.Delta == 1.0) { + + // Compute the Ix2_infty numbers + + int sum1 = 0; + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + sum1 = 0; + + for (int k = 0; k < RefChain.Nstrings; ++k) { + + sum1 += Nrap[k] * (2 * JSC::min(RefChain.Str_L[j], RefChain.Str_L[k]) - ((j == k) ? 1 : 0)); + } + + //Ix2_infty[j] = (RefChain.Nsites - 1.0 + 2.0 * RefChain.Str_L[j] - sum1); + Ix2_infty[j] = (RefChain.Nsites + 1.0 - sum1); // to get counting right... + + } // The Ix2_infty are now set. + + // Now compute the Ix2_max limits + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + Ix2_max[j] = int(floor(Ix2_infty[j])); // sets basic integer + + // Give the correct parity to Ix2_max + + // If Nrap is even, Ix2_max must be odd. If odd, then even. + + if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; + + // If Ix2_max equals Ix2_infty, we reduce it by 2: + + if (Ix2_max[j] == int(Ix2_infty[j])) Ix2_max[j] -= 2; + + while (Ix2_max[j] > RefChain.Nsites) { + Ix2_max[j] -= 2; + } + } + + } // if XXX AFM + + else if (RefChain.Delta > 1.0) { + + // Compute the Ix2_infty numbers + + int sum1 = 0; + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + sum1 = 0; + + for (int k = 0; k < RefChain.Nstrings; ++k) { + + sum1 += Nrap[k] * (2 * JSC::min(RefChain.Str_L[j], RefChain.Str_L[k]) - ((j == k) ? 1 : 0)); + } + + Ix2_infty[j] = (RefChain.Nsites - 1 + 2 * RefChain.Str_L[j] - sum1); + + } // The Ix2_infty are now set. + + // Now compute the Ix2_max limits + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + Ix2_max[j] = int(floor(Ix2_infty[j])); // sets basic integer + + // Give the correct parity to Ix2_max + + // If Nrap is even, Ix2_max must be odd. If odd, then even. + + if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; + + // If Ix2_max equals Ix2_infty, we reduce it by 2: + + //if (Ix2_max[j] == Ix2_infty[j]) Ix2_max[j] -= 2; + + while (Ix2_max[j] > RefChain.Nsites) { + Ix2_max[j] -= 2; + } + + // Fudge, for strings: + //if (RefChain.Str_L[j] >= 1) Ix2_max[j] += 2; + //Ix2_max[j] += 2; + } + + } // if XXZ_gpd + + } + + + + +} // namespace JSC diff --git a/src/BETHE/Bethe_State.cc b/src/BETHE/Bethe_State.cc new file mode 100644 index 0000000..c0846d4 --- /dev/null +++ b/src/BETHE/Bethe_State.cc @@ -0,0 +1,29 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/BETHE/Bethe_State.cc + +Purpose: defines functions in Bethe_State class, + providing a unified object for eigenstates of all + Bethe Ansatz integrable models. + + + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + Bethe_State::Bethe_State (long long int base_id_ref, long long int type_id_ref, long long int id_ref, long long int maxid_ref) : + base_id(base_id_ref), type_id(type_id_ref), id(id_ref), maxid(maxid_ref) {} + +} // namespace JSC diff --git a/src/BETHE/Offsets.cc b/src/BETHE/Offsets.cc new file mode 100644 index 0000000..beaf0b8 --- /dev/null +++ b/src/BETHE/Offsets.cc @@ -0,0 +1,277 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) 2006-9. + +----------------------------------------------------------- + +File: src/BETHE/Offsets.cc + +Purpose: defines functions in Offsets class. + + +Last modified: 19/10/2009 + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + // Function definitions: class Offsets + + Offsets::Offsets () : base(), Tableau(Vect()), type_id(0LL), id(0LL), maxid(0LL) {}; + + Offsets::Offsets (const Offsets& RefOffset) // copy constructor + : base(RefOffset.base), Tableau(Vect (2 * base.Nrap.size() + 2)), type_id(RefOffset.type_id), id(RefOffset.id), maxid(RefOffset.maxid) + { + for (int i = 0; i < 2 * base.Nrap.size() + 2; ++i) Tableau[i] = RefOffset.Tableau[i]; + } + + Offsets::Offsets (const Heis_Base& RefBase, long long int req_type_id) + // sets all tableaux to empty ones, with nparticles(req_type_id) at each level + { + // Build nparticles vector from req_type_id + + Vect nparticles(0, 2* RefBase.Nrap.size() + 2); + long long int factor = pow_ulli (10LL, nparticles.size() - 1); + long long int id_eff = req_type_id; + for (int i = 0; i < nparticles.size(); ++i) { + nparticles[nparticles.size() - 1 - i] = id_eff/factor; + id_eff -= factor * nparticles[nparticles.size() - 1 - i]; + factor /= 10LL; + } + + // Check if we've got the right vector... + long long int idcheck = Offsets_type_id (nparticles); + if (idcheck != req_type_id) JSCerror("idcheck != req_type_id in Offsets constructor."); + + (*this) = Offsets(RefBase, nparticles); + } + + Offsets::Offsets (const Heis_Base& RefBase, Vect nparticles) // sets all tableaux to empty ones, with nparticles at each level + : base(RefBase), Tableau(Vect (2 * base.Nrap.size() + 2)), type_id(Offsets_type_id (nparticles)), id(0LL), maxid(0LL) + { + + // Checks on nparticles vector: + + if (nparticles.size() != 2 * base.Nrap.size() + 2) JSCerror("Wrong nparticles.size in Offsets constructor."); + + //if (base.Nrap[0] != (nparticles[3] + nparticles[2] + base.Mdown - nparticles[0] - nparticles[1])) JSCerror("Wrong Nrap[0] in Offsets constructor."); + if (nparticles[3] + nparticles[2] != nparticles[0] + nparticles[1]) { + cout << nparticles[0] << "\t" << nparticles[1] << "\t" << nparticles[2] << "\t" << nparticles[3] << endl; + JSCerror("Wrong Npar[0-3] in Offsets constructor."); + } + + for (int base_level = 1; base_level < base.Nrap.size(); ++ base_level) + if (base.Nrap[base_level] != nparticles[2*base_level + 2] + nparticles[2*base_level + 3]) { + cout << base_level << "\t" << base.Nrap[base_level] << "\t" << nparticles[2*base_level + 2] << "\t" << nparticles[2*base_level + 3] << endl; + JSCerror("Wrong Nrap[] in Offsets constructor."); + } + + // nparticles[0,1]: number of holes on R and L side in GS interval + if (nparticles[0] > (base.Nrap[0] + 1)/2) JSCerror("nparticles[0] too large in Offsets constructor."); + if (nparticles[1] > base.Nrap[0]/2) JSCerror("nparticles[1] too large in Offsets constructor."); + + // nparticles[2,3]: number of particles of type 0 on R and L side out of GS interval + if (nparticles[2] > (base.Ix2_max[0] - base.Nrap[0] + 1)/2) JSCerror("nparticles[2] too large in Offsets constructor."); + if (nparticles[3] > (base.Ix2_max[0] - base.Nrap[0] + 1)/2) JSCerror("nparticles[3] too large in Offsets constructor."); + + for (int base_level = 1; base_level < base.Nrap.size(); ++ base_level) + if ((nparticles[2*base_level + 2] > 0 && nparticles[2*base_level + 2] > (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2) + //|| (nparticles[2*base_level + 3] > 0 && nparticles[2*base_level + 3] > (base.Ix2_max[base_level] - (base.Nrap[base_level] % 2) - 1)/2)) { + || (nparticles[2*base_level + 3] > 0 + && nparticles[2*base_level + 3] > base.Ix2_max[base_level] + 1 - (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2)) { + cout << base_level << "\t" << nparticles[2*base_level + 2] << "\t" << (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2 + << "\t" << nparticles[2*base_level + 3] << "\t" << (base.Ix2_max[base_level] - (base.Nrap[base_level] % 2) - 1)/2 + << "\t" << (nparticles[2*base_level + 2] > 0) << "\t" << (nparticles[2*base_level + 2] > (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2) + //<< "\t" << (nparticles[2*base_level + 3] > 0) << "\t" << (nparticles[2*base_level + 3] > (base.Ix2_max[base_level] - (base.Nrap[base_level] % 2) - 1)/2) + << "\t" << (nparticles[2*base_level + 3] > 0) << "\t" + << (nparticles[2*base_level + 3] > base.Ix2_max[base_level] + 1 - (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2) + << endl; + JSCerror("nparticles too large in Offsets constructor."); + } + + // Check sum of rapidities + + // Holes in GS interval + Tableau[0] = Young_Tableau(nparticles[0], (base.Nrap[0] + 1)/2 - nparticles[0]); + Tableau[1] = Young_Tableau(nparticles[1], base.Nrap[0]/2 - nparticles[1], Tableau[0]); + + // Particles of type 0 out of GS interval + Tableau[2] = Young_Tableau(nparticles[2], (base.Ix2_max[0] - base.Nrap[0] + 1)/2 - nparticles[2], Tableau[0]); + Tableau[3] = Young_Tableau(nparticles[3], (base.Ix2_max[0] - base.Nrap[0] + 1)/2 - nparticles[3], Tableau[2]); + + // Tableaux of index i = 2,...: data about string type i/2-1. + for (int base_level = 1; base_level < base.Nrap.size(); ++base_level) { + Tableau[2*base_level + 2] = Young_Tableau(nparticles[2*base_level + 2], + //(base.Ix2_max[base_level] - ((base.Nrap[base_level]) % 2) + 2)/2 - nparticles[2*base_level + 2], Tableau[2]); + //(base.Ix2_max[base_level] - base.Nrap[base_level] % 2 + 2)/2 - nparticles[2*base_level + 2], Tableau[2]); + (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2))/2 + 1 - nparticles[2*base_level + 2], Tableau[2]); + Tableau[2*base_level + 3] = Young_Tableau(nparticles[2*base_level + 3], + //(base.Ix2_max[base_level] - base.Nrap[base_level] % 2)/2 - nparticles[2*base_level + 3], Tableau[3]); + (base.Ix2_max[base_level] - (base.Nrap[base_level] % 2) - 1)/2 + 1 - nparticles[2*base_level + 3], Tableau[3]); + } + + maxid = 1LL; + //id = Tableau[0].id; + for (int i = 0; i < nparticles.size(); ++i) { + maxid *= Tableau[i].maxid + 1LL; + //id += maxid + Tableau[i].id; + } + maxid -= 1LL; + + } + + Offsets& Offsets::operator= (const Offsets& RefOffset) + { + if (this != &RefOffset) { + base = RefOffset.base; + Tableau = RefOffset.Tableau; + type_id = RefOffset.type_id; + id = RefOffset.id; + maxid = RefOffset.maxid; + } + return(*this); + } + + bool Offsets::operator<= (const Offsets& RefOffsets) + { + // Check whether all nonzero tableau row lengths in RefOffsets + // are <= than those in *this + + bool answer = true; + for (int level = 0; level < 4; ++level) { // check fundamental level only + //for (int level = 0; level < 2 * base.Nrap.size() + 2; ++level) { + // First check whether all rows which exist in both tableaux satisfy rule: + for (int tableau_level = 0; tableau_level < JSC::min(Tableau[level].Nrows, RefOffsets.Tableau[level].Nrows); ++tableau_level) + if (Tableau[level].Row_L[tableau_level] > RefOffsets.Tableau[level].Row_L[tableau_level]) + answer = false; + // Now check whether there exist extra rows violating rule: + for (int tableau_level = JSC::min(Tableau[level].Nrows, RefOffsets.Tableau[level].Nrows); tableau_level < Tableau[level].Nrows; ++tableau_level) + if (Tableau[level].Row_L[tableau_level] > 0) answer = false; + } + + return(answer); + } + + bool Offsets::operator>= (const Offsets& RefOffsets) + { + // Check whether all nonzero tableau row lengths in RefOffsets + // are >= than those in *this + + bool answer = true; + for (int level = 0; level < 4; ++level) { // check fundamental level only + //for (int level = 0; level < 2 * base.Nrap.size() + 2; ++level) { + // First check whether all rows which exist in both tableaux satisfy rule: + for (int tableau_level = 0; tableau_level < JSC::min(Tableau[level].Nrows, RefOffsets.Tableau[level].Nrows); ++tableau_level) + if (Tableau[level].Row_L[tableau_level] < RefOffsets.Tableau[level].Row_L[tableau_level]) + answer = false; + // Now check whether there exist extra rows violating rule: + for (int tableau_level = JSC::min(Tableau[level].Nrows, RefOffsets.Tableau[level].Nrows); tableau_level < RefOffsets.Tableau[level].Nrows; ++tableau_level) + if (RefOffsets.Tableau[level].Row_L[tableau_level] > 0) answer = false; + } + + return(answer); + } + + void Offsets::Compute_type_id () + { + type_id = 0LL; + + for (int i = 0; i < 2*base.Nrap.size() + 2; ++i) { + Tableau[i].Compute_id(); + type_id += Tableau[i].Nrows * pow_ulli(10LL, i); + } + } + + void Offsets::Set_to_id (long long int idnr) + { + // The idnr of the Offset is given by + // sub_id[0] + (total number of tableaux of type 0) * (sub_id[1] + (total number of tableaux of type 1) * (sub_id[2] + ... + // + total number of tableaux of type (2*base.Nrap.size()) * sub_id[2*base.Nrap.size() + 1] + + if (idnr > maxid) { + cout << idnr << "\t" << maxid << endl; + JSCerror("idnr too large in offsets.Set_to_id."); + } + + id = idnr; + + Vect sub_id(0LL, 2*base.Nrap.size() + 2); + + long long int idnr_eff = idnr; + long long int temp_prod = 1LL; + + Vect result_choose(2*base.Nrap.size() + 2); + + for (int i = 0; i <= 2*base.Nrap.size(); ++i) { + //result_choose[i] = choose_lli(Tableau[i].Nrows + Tableau[i].Ncols, Tableau[i].Nrows); + result_choose[i] = Tableau[i].maxid + 1LL; + temp_prod *= result_choose[i]; + } + + for (int i = 2*base.Nrap.size() + 1; i > 0; --i) { + sub_id[i] = idnr_eff/temp_prod; + idnr_eff -= sub_id[i] * temp_prod; + temp_prod /= result_choose[i-1]; + } + sub_id[0] = idnr_eff; // what's left goes to the bottom... + + for (int i = 0; i <= 2*base.Nrap.size() + 1; ++i) { + //cout << "level = " << i << " Tableau.id = " << sub_id[i] << endl; + if ((Tableau[i].Nrows * Tableau[i].Ncols == 0) && (sub_id[i] != 0)) JSCerror("index too large in offset.Set_to_id."); + if (Tableau[i].id != sub_id[i]) Tableau[i].Set_to_id(sub_id[i]); + } + + Compute_type_id (); + + return; + } + + void Offsets::Compute_id () + { + long long int prod_maxid = 1LL; + + id = 0LL; + + for (int i = 0; i < 2*base.Nrap.size() + 2; ++i) { + Tableau[i].Compute_id(); + id += Tableau[i].id * prod_maxid; + prod_maxid *= Tableau[i].maxid + 1LL; + } + } + + Vect Offsets::Descendents (bool fixed_iK) + { + // From a given vector of Young tableaux specifying a particular eigenstate, + // this function provides the full set of descendents (either at the same momentum if + // fixed_iK == true, or not) by returning a vector of all descendent id's (leaving the + // base and type invariant), which can then be used for further calculations. + + // This set of descendents is meant to be used when calculating either partition functions + // or zero-temperature correlation functions. + + // IMPORTANT ASSUMPTIONS: + // - all even sectors consistently increase/decrease momentum for increasing tableau row length + // - all odd sectors consistently decrease/increase momentum for increasing tableau row length + + // FOR FIXED MOMENTUM: + // all tableau levels `above' the lowest occupied one are descended as for fixed_iK == false, + // and the lowest sector's highest tableau level's row length is modified (increased or decreased by one + // unit if possible) such that the iK of Tableau_desc == iK of Tableau_ref. + // The logic behind this is that for a state with nexc excitations, we let run nexc - 1 of the + // excitations, and the lowest one is fixed in place by the momentum constraint, if possible. + + Vect Tableau_ref = (*this).Tableau; + Vect Tableau_desc = Tableau_ref; + + + } + + + +} // namespace JSC diff --git a/src/COMBI/Combinatorics.cc b/src/COMBI/Combinatorics.cc new file mode 100644 index 0000000..d1257fd --- /dev/null +++ b/src/COMBI/Combinatorics.cc @@ -0,0 +1,89 @@ +/**************************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) 2006. + +----------------------------------------------------------- + +Combinatorics.cc + +Defines all class related to combinatorics. + +LAST MODIFIED: 04/09/06 + +******************************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + Choose_Table::Choose_Table () + : Npower(0ULL), Npowerp1(1ULL), table(new unsigned long long int[1]) + { + table[0] = 1ULL; + } + + Choose_Table::Choose_Table (int Npower_ref) + : Npower(Npower_ref), Npowerp1(Npower_ref + 1ULL) + { + dim = Npowerp1 * Npowerp1; + + // We can only go up to ULL_MAX: + if (log(DP(ULONG_LONG_MAX)) < DP(Npowerp1) * log(2.0)) + JSCerror("Choose_Table: too large to contruct."); + + table = new unsigned long long int[dim]; + + (*this).Fill_table(); + } + + void Choose_Table::Fill_table() + { + table[0] = 1ULL; + int n,m; + for (n = 0; n <= Npower; ++n) { + table[Npowerp1 * n] = 1ULL; + for (m = 1; m < n; ++m) { + table[Npowerp1 * n + m] = table[Npowerp1 * (n-1) + m - 1] + table[Npowerp1 * (n-1) + m]; + } + table[Npowerp1 * n + n] = 1ULL; + for (m = n+1; m <= Npower; ++m) + table[Npowerp1 * n + m] = 0ULL; + } + } + + int Choose_Table::power() + { + return(Npower); + } + + unsigned long long int Choose_Table::choose(int N, int M) + { + if (N < 0 || N > Npower) JSCerror("N out of bounds in choose(N,M)."); + if (M < 0 || M > Npower) JSCerror("M out of bounds in choose(N,M)."); + + return(table[Npowerp1 * N + M]); + } + + std::ostream& operator<< (std::ostream& s, Choose_Table& Ref_table) + { + s << endl; + for (int n = 0; n <= Ref_table.power(); ++n) { + for (int m = 0; m <= Ref_table.power(); ++m) + s << Ref_table.choose(n, m) << " "; + s << endl; + } + s << endl; + return(s); + } + + Choose_Table::~Choose_Table() + { + delete[] table; + } + + +} // namespace JSC diff --git a/src/EXECS/2CBG_ThLim.cc b/src/EXECS/2CBG_ThLim.cc new file mode 100755 index 0000000..006b708 --- /dev/null +++ b/src/EXECS/2CBG_ThLim.cc @@ -0,0 +1,43 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: 2CBG_ThLim.cc + +Purpose: solves the TBA equations for the 2-component Bose gas + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +int main(int argc, const char* argv[]) +{ + + if (argc != 7) JSCerror("Wrong number of arguments to 2CBG_ThLim executable. Use c(best to set to 1), mu, Omega, kBT, TT(minutes), bool Save_data (0 == false)."); + + DP c_int = atof(argv[1]); + DP mu = atof(argv[2]); + DP Omega = atof(argv[3]); + DP kBT = atof(argv[4]); + int Max_Secs = 60 * atoi(argv[5]); + bool Save_data = bool(atoi(argv[6])); + + if (c_int <= 0.0) JSCerror("Give a strictly positive c."); + if (Omega <= 0.0) JSCerror("Give a strictly positive Omega, otherwise the algorithm cannot converge."); + if (kBT <= 0.0) JSCerror("Negative T ? You must be a string theorist."); + if (Max_Secs < 10) JSCerror("Give more time."); + + //cout << "Read c_int = " << c_int << "\tmu = " << mu << "\tOmega = " << Omega << "\tkBT = " << kBT << "\tMax_Secs = " << Max_Secs << endl; + + Solve_2CBG_TBAE_via_refinements (c_int, mu, Omega, kBT, Max_Secs, Save_data); + + return(0); +} diff --git a/src/EXECS/Analyze_RAW_File.cc b/src/EXECS/Analyze_RAW_File.cc new file mode 100644 index 0000000..d8d6557 --- /dev/null +++ b/src/EXECS/Analyze_RAW_File.cc @@ -0,0 +1,84 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Analyze_RAW_File.cc + +Purpose: give some statistics for the matrix element distribution in a raw file. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 4) { + cout << "Argument needed: rawfile, iKmin, iKmax." << endl; + JSCerror(""); + } + + const char* rawfilename = argv[1]; + int iKmin = atoi(argv[2]); + int iKmax = atoi(argv[3]); + + ifstream RAW_infile; + RAW_infile.open(rawfilename); + if (RAW_infile.fail()) { + cout << rawfilename << endl; + JSCerror("Could not open RAW_infile... "); + } + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + DP sumFFsq = 0.0; + DP sumFF4 = 0.0; + DP sumFFsqlnFFsq = 0.0; + + Vect nFFatK (0, iKmax - iKmin + 1); + Vect sumFFsqatK (0.0, iKmax - iKmin + 1); + Vect sumFF4atK (0.0, iKmax - iKmin + 1); + Vect sumFFsqlnFFsqatK (0.0, iKmax - iKmin + 1); + + int nread = 0; + + while (RAW_infile.peek() != EOF) { + RAW_infile >> omega >> iK >> FF >> dev >> label; + + nread++; + sumFFsq += FF*FF; + sumFF4 += FF*FF*FF*FF; + sumFFsqlnFFsq += FF*FF * log(FF*FF); + + if (iK >= iKmin && iK <= iKmax) { + nFFatK[iK-iKmin] += 1; + sumFFsqatK[iK - iKmin] += FF*FF; + sumFF4atK[iK - iKmin] += FF*FF*FF*FF; + sumFFsqlnFFsqatK[iK - iKmin] += FF*FF * log(FF*FF); + } + } + + RAW_infile.close(); + + cout << "Inverse participation ratio: \t" << sumFF4/(sumFFsq*sumFFsq) << endl; + // Entropy is -sum (FFsq/sumFFsq) * ln(FFsq/sumFFsq) = sum (FFsq lnFFsq - FFsq ln sumFFsq)/sumFFsq + cout << "Entropy: \t" << -(sumFFsqlnFFsq - sumFFsq * log(sumFFsq))/sumFFsq << endl; + + cout << "iK\tnFFatK\tIPRatiK\tentropyatiK:" << endl; + for (int iK = iKmin; iK <= iKmax; ++iK) cout << iK << "\t" << nFFatK[iK-iKmin] << "\t" << sumFF4atK[iK-iKmin]/(sumFFsqatK[iK - iKmin] * sumFFsqatK[iK - iKmin]) << "\t" << -(sumFFsqlnFFsqatK[iK-iKmin] - sumFFsqatK[iK-iKmin] * log(sumFFsqatK[iK-iKmin]))/sumFFsqatK[iK-iKmin]<< endl; + + + return(0); +} diff --git a/src/EXECS/Check_RAW_File.cc b/src/EXECS/Check_RAW_File.cc new file mode 100644 index 0000000..62939b9 --- /dev/null +++ b/src/EXECS/Check_RAW_File.cc @@ -0,0 +1,147 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Check_RAW_File.cc + +Purpose: from a .raw_srt file, check that nonzero momentum states + appear (only) twice, and zero momentum ones (only) once. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 7) { // print out some instructions + cout << "Usage of Check_RAW_File executable: provide the following arguments:" << endl; + cout << "(sorted!) raw file name, iKmin, iKmax, sympoint, FFmin, check_option." << endl; + cout << "Check option: 0 == check for missing states, 1 == check for multiply-appearing states." << endl; + } + + char* rawfilename = argv[1]; + int iKmin = atoi(argv[2]); + int iKmax = atoi(argv[3]); + int sympoint = atoi(argv[4]); + DP FFmin = atof(argv[5]); + int check_option = atoi(argv[6]); + + ifstream RAW_infile; + RAW_infile.open(rawfilename); + if (RAW_infile.fail()) { + cout << rawfilename << endl; + JSCerror("Could not open sorted RAW_infile... "); + } + + DP omega_next, omega, omega_prev; + int iK_next, iK, iK_prev; + DP FF_next, FF, FF_prev; + //int conv_next, conv, conv_prev; + DP dev_next, dev, dev_prev; + string label_next, label, label_prev; + + if (check_option > 1) { + FF = 0.0; + FF_prev = 0.0; + FF_next = 0.0; + FFmin = -1.0; + } + + //RAW_infile >> omega >> iK >> FF >> conv >> label; + RAW_infile >> omega >> iK; + if (check_option <= 1) RAW_infile >> FF; + RAW_infile >> dev; + RAW_infile >> label; + //RAW_infile >> omega_next >> iK_next >> FF_next >> conv_next >> label_next; + RAW_infile >> omega_next >> iK_next; + if (check_option <= 1) RAW_infile >> FF_next; + RAW_infile >> dev_next; + RAW_infile >> label_next; + + int line = 1; + + char a; + + + while (fabs(FF) > FFmin && RAW_infile.peek() != EOF) { + + //omega_prev = omega; iK_prev = iK; FF_prev = FF; conv_prev = conv; label_prev = label; + omega_prev = omega; iK_prev = iK; FF_prev = FF; dev_prev = dev; label_prev = label; + //omega = omega_next; iK = iK_next; FF = FF_next; conv = conv_next; label = label_next; + omega = omega_next; iK = iK_next; FF = FF_next; dev = dev_next; label = label_next; + //RAW_infile >> omega_next >> iK_next >> FF_next >> conv_next >> label_next; + RAW_infile >> omega_next >> iK_next; + if (check_option <= 1) RAW_infile >> FF_next; // for non-Z checks + RAW_infile >> dev_next; + RAW_infile >> label_next; + //cout << "checking line " << line << endl; + //cout << omega_prev << "\t" << iK_prev << "\t" << FF_prev << "\t" << label_prev << endl + // << omega << "\t" << iK << "\t" << FF << "\t" << label << endl + // << omega_next << "\t" << iK_next << "\t" << FF_next << "\t" << label_next << endl; + line++; + + if (label.compare(label_next) == 0) + cout << "Identical labels around line " << line << ": " << endl + << omega << "\t" << iK << "\t" << FF << "\t" << dev << "\t" << label << endl; + + if (check_option == 0 && iK != 0 && iK != sympoint && iK >= iKmin && iK <= iKmax + && fabs((FF - FF_prev)/(FF + FF_prev)) > 1.0e-6 && fabs((FF - FF_next)/(FF + FF_next)) > 1.0e-6) { + + cout << "State missing around line " << line << ": " << endl + //<< omega_prev << "\t" << iK_prev << "\t" << FF_prev << "\t" << conv_prev << "\t" << label_prev << endl + << omega_prev << "\t" << iK_prev << "\t" << FF_prev << "\t" << dev_prev << "\t" << label_prev << endl + //<< omega << "\t" << iK << "\t" << FF << "\t" << conv << "\t" << label << endl + << omega << "\t" << iK << "\t" << FF << "\t" << dev << "\t" << label << endl + //<< omega_next << "\t" << iK_next << "\t" << FF_next << "\t" << conv_next << "\t" << label_next << endl; + << omega_next << "\t" << iK_next << "\t" << FF_next << "\t" << dev_next << "\t" << label_next << endl; + + cin >> a; + //break; + } + + if (check_option == 1 && iK_prev == iK && iK == iK_next && fabs((omega - omega_prev)/(omega + omega_prev)) < 1.0e-8 && fabs((omega - omega_next)/(omega + omega_next)) < 1.0e-8 && fabs((FF - FF_prev)/(FF + FF_prev)) < 1.0e-8 && fabs((FF - FF_next)/(FF + FF_next)) < 1.0e-8) { + + cout << "Triple state around line " << line << ": " << endl + //<< omega_prev << "\t" << iK_prev << "\t" << FF_prev << "\t" << conv_prev << "\t" << label_prev << endl + << omega_prev << "\t" << iK_prev << "\t" << FF_prev << "\t" << dev_prev << "\t" << label_prev << endl + //<< omega << "\t" << iK << "\t" << FF << "\t" << conv << "\t" << label << endl + << omega << "\t" << iK << "\t" << FF << "\t" << dev << "\t" << label << endl + //<< omega_next << "\t" << iK_next << "\t" << FF_next << "\t" << conv_next << "\t" << label_next << endl; + << omega_next << "\t" << iK_next << "\t" << FF_next << "\t" << dev_next << "\t" << label_next << endl; + cin >> a; + } + + if (check_option == 2 && iK != 0 && iK != sympoint && iK >= iKmin && iK <= iKmax + && fabs((omega - omega_prev)/(omega + omega_prev)) > 1.0e-6 && fabs((omega - omega_next)/(omega + omega_next)) > 1.0e-6) { + + cout << "State missing around line " << line << ": " << endl + << omega_prev << "\t" << iK_prev << "\t" << dev_prev << "\t" << label_prev << endl + << omega << "\t" << iK << "\t" << dev << "\t" << label << endl + << omega_next << "\t" << iK_next << "\t" << dev_next << "\t" << label_next << endl; + + cin >> a; + //break; + } + + if (check_option == 3 && iK_prev == iK && iK == iK_next && fabs((omega - omega_prev)/(omega + omega_prev)) < 1.0e-8 && fabs((omega - omega_next)/(omega + omega_next)) < 1.0e-8) { + + cout << "Triple state around line " << line << ": " << endl + << omega_prev << "\t" << iK_prev << "\t" << dev_prev << "\t" << label_prev << endl + << omega << "\t" << iK << "\t" << dev << "\t" << label << endl + << omega_next << "\t" << iK_next << "\t" << dev_next << "\t" << label_next << endl; + cin >> a; + } + + } + + return(0); +} + diff --git a/src/EXECS/Heis_DSF.cc b/src/EXECS/Heis_DSF.cc new file mode 100644 index 0000000..04c4832 --- /dev/null +++ b/src/EXECS/Heis_DSF.cc @@ -0,0 +1,142 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Heis_DSF.cc + +Purpose: main function for ABACUS++ for Heisenberg spin-1/2 chain + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 8) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Heis_DSF executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: m for S- S+, z for Sz Sz, p for S+ S-." << endl; + cout << "DP Delta \t\t Value of the anisotropy: use positive real values only" << endl; + cout << "int N \t\t\t Length (number of sites) of the system: use positive even integer values only" << endl; + cout << "int M \t\t\t Number of down spins: use positive integer values between 1 and N/2" << endl; + //cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: 0 and N" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "bool refine \t\t Is this a refinement of a earlier calculations ? (0 == false, 1 == true)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "Heis_DSF z 1.0 100 40 0 100 600 1.0 0" << endl << endl; + } + + else if (argc == 8) { // !fixed_iK + int ctr = 1; + char whichDSF = *argv[ctr++]; + DP Delta = atof(argv[ctr++]); + int N = atoi(argv[ctr++]); + int M = atoi(argv[ctr++]); + //int iKmin = atoi(argv[5]); + //int iKmax = atoi(argv[6]); + int Max_Secs = atoi(argv[ctr++]); + DP target_sumrule = atof(argv[ctr++]); + bool refine = (atoi(argv[ctr++]) == 1); + + // We systematically scan over all momentum integers (to avoid problems with Brillouin folding + int iKmin = -1000*N; + int iKmax = 1000*N; + //Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, Max_Secs, target_sumrule, refine, 0, 1); + Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, Max_Secs, target_sumrule, refine); + } + + + + // The argument given is the name of the standard args_Heis_DSF arguments file + + /* + if (argc == 2) { // Used an input file to provide the arguments + + if (strcmp(argv[1],"help") == 0) { // Output some instructions + cout << "Usage of Heis_DSF executable: " << endl; + cout << endl << "Provide arguments by either using one of the three following options:" << endl << endl; + cout << "1) via an argument file (see the template `args_Heis_DSF' in directory src/EXECS/), for example" << endl << endl; + cout << "Heis_DSF args_Heis_DSF" << endl << endl; + cout << "2) with arguments (for general momenta scan) whichDSF Delta N M iKmin iKmax Max_Secs refine, for example" << endl << endl; + cout << "Heis_DSF z 0.9 100 40 0 50 600 0" << endl << endl; + cout << "3) with arguments (for general momenta scan) whichDSF Delta N M iKneeded Max_Secs refine, for example" << endl << endl; + cout << "Heis_DSF z 0.9 100 40 20 600 0" << endl << endl; + } + + else { // read argument file + + ifstream argsfile; + argsfile.open(argv[1]); + if (argsfile.fail()) { + cout << argv[1] << endl; + JSCerror("Could not open arguments file."); + } + + char junk[256]; + while (argsfile.peek() == '#' || argsfile.peek() == '\t' || argsfile.peek() == ' ' || argsfile.peek() == '\n') argsfile.getline(junk, 256); + char whichDSF; argsfile >> whichDSF; + while (argsfile.peek() == '#' || argsfile.peek() == '\t' || argsfile.peek() == ' ' || argsfile.peek() == '\n') argsfile.getline(junk, 256); + DP Delta; argsfile >> Delta; + while (argsfile.peek() == '#' || argsfile.peek() == '\t' || argsfile.peek() == ' ' || argsfile.peek() == '\n') argsfile.getline(junk, 256); + int N; argsfile >> N; + while (argsfile.peek() == '#' || argsfile.peek() == '\t' || argsfile.peek() == ' ' || argsfile.peek() == '\n') argsfile.getline(junk, 256); + int M; argsfile >> M; + while (argsfile.peek() == '#' || argsfile.peek() == '\t' || argsfile.peek() == ' ' || argsfile.peek() == '\n') argsfile.getline(junk, 256); + //bool fixed_iK; argsfile >> fixed_iK; + //while (argsfile.peek() == '#' || argsfile.peek() == '\t' || argsfile.peek() == ' ' || argsfile.peek() == '\n') argsfile.getline(junk, 256); + //int iKneeded; argsfile >> iKneeded; + int iKmin, iKmax; argsfile >> iKmin >> iKmax; + while (argsfile.peek() == '#' || argsfile.peek() == '\t' || argsfile.peek() == ' ' || argsfile.peek() == '\n') argsfile.getline(junk, 256); + int Max_Secs; argsfile >> Max_Secs; + while (argsfile.peek() == '#' || argsfile.peek() == '\t' || argsfile.peek() == ' ' || argsfile.peek() == '\n') argsfile.getline(junk, 256); + bool refine; argsfile >> refine; + + Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, Max_Secs, refine); + } + } // if (argc == 2) + + else if (argc == 8) { // fixed_iK + char whichDSF = *argv[1]; + DP Delta = atof(argv[2]); + int N = atoi(argv[3]); + int M = atoi(argv[4]); + int iKneeded = atoi(argv[5]); + int Max_Secs = atoi(argv[6]); + bool refine = (atoi(argv[7]) == 1); + + //Scan_Heis (whichDSF, Delta, N, M, iKneeded, Max_Secs, refine); + Scan_Heis (whichDSF, Delta, N, M, iKneeded, iKneeded, Max_Secs, refine); + } + + else if (argc == 9) { // !fixed_iK + char whichDSF = *argv[1]; + DP Delta = atof(argv[2]); + int N = atoi(argv[3]); + int M = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + int Max_Secs = atoi(argv[7]); + bool refine = (atoi(argv[8]) == 1); + + //Scan_Heis (whichDSF, Delta, N, M, iKneeded, Max_Secs, refine); + Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, Max_Secs, refine); + } + + else JSCerror("Wrong number of arguments to Heis_DSF executable."); + */ + + return(0); +} + diff --git a/src/EXECS/Heis_DSF_GeneralState.cc b/src/EXECS/Heis_DSF_GeneralState.cc new file mode 100644 index 0000000..c87667b --- /dev/null +++ b/src/EXECS/Heis_DSF_GeneralState.cc @@ -0,0 +1,133 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Heis_DSF_GeneralState.cc + +Purpose: main function for ABACUS++ for Heisenberg spin-1/2 chain + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 9) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Heis_DSF executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: m for S- S+, z for Sz Sz, p for S+ S-." << endl; + cout << "DP Delta \t\t Value of the anisotropy: use positive real values only" << endl; + cout << "int N \t\t\t Length (number of sites) of the system: use positive even integer values only" << endl; + cout << "int M \t\t\t Number of down spins: use positive integer values between 1 and N/2" << endl; + cout << "char* defaultScanStatename:\t\t file [].Ix2 contains the quantum numbers defining the AveragingState; used as defaultScanStatename" << endl; + //cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: 0 and N" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "bool refine \t\t Is this a refinement of a earlier calculations ? (0 == false, 1 == true)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + } + + else if (argc == 9) { // !fixed_iK + int ctr = 1; + char whichDSF = *argv[ctr++]; + DP Delta = atof(argv[ctr++]); + int N = atoi(argv[ctr++]); + int M = atoi(argv[ctr++]); + char* Ix2filenameprefix = argv[ctr++]; + //int iKmin = atoi(argv[5]); + //int iKmax = atoi(argv[6]); + int Max_Secs = atoi(argv[ctr++]); + DP target_sumrule = atof(argv[ctr++]); + bool refine = (atoi(argv[ctr++]) == 1); + + // We systematically scan over all momentum integers (to avoid problems with Brillouin folding + int iKmin = -1000*N; + int iKmax = 1000*N; + + // Read the Ix2 from the file: + // Format is: + // base_level, Nrap[base_level], \endl Ix2[base_level], repeat for all occupied base_levels... + ifstream Ix2_input_file; + stringstream filenamestrstream; + filenamestrstream << Ix2filenameprefix; + string defaultScanStatename = filenamestrstream.str(); + filenamestrstream << ".Ix2"; + string filenamestr = filenamestrstream.str(); + const char* filename_Cstr = filenamestr.c_str(); + Ix2_input_file.open(filename_Cstr); + if (Ix2_input_file.fail()) { + cout << filename_Cstr << endl; + JSCerror("Could not open Ix2 input file in Heis_DSF_GeneralState"); + } + + Heis_Chain chain(1.0, Delta, 0.0, N); + + Vect Nrap_read (0, chain.Nstrings); + int level = 0; + Vect > Ix2_read (chain.Nstrings); + do { + Ix2_input_file >> level; + Ix2_input_file >> Nrap_read[level]; + Ix2_read[level] = Vect (Nrap_read[level]); + for (int alpha = 0; alpha < Nrap_read[level]; ++alpha) Ix2_input_file >> Ix2_read[level][alpha]; + //cout << "Read level = " << level << "\tNrap_read[level] = " << Nrap_read[level] << endl; + //cout << "\tIx2_read[level] = " << Ix2_read[level] << endl; + } while (Ix2_input_file.peek() != EOF); + + // Construct the Averaging State: + Heis_Base base (chain, Nrap_read); + + int paralevel = 0; + Vect rank; // dummy + Vect nr_processors; // dummy + + if (Delta > 0.0 && Delta < 1.0) { + XXZ_Bethe_State AveragingState (chain, base); + for (int il = 0; il < chain.Nstrings; ++il) { + if (Nrap_read[il] > 0) for (int alpha = 0; alpha < Nrap_read[il]; ++alpha) AveragingState.Ix2[il][alpha] = Ix2_read[il][alpha]; + } + AveragingState.Set_Label_from_Ix2(AveragingState.Ix2); + AveragingState.Compute_All(true); + + //cout << "AveragingState read from file: " << AveragingState << endl; + // Perform the scan: + Scan_Heis (whichDSF, AveragingState, defaultScanStatename, iKmin, iKmax, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + } + else if (Delta == 1.0) { + XXX_Bethe_State AveragingState (chain, base); + for (int il = 0; il < chain.Nstrings; ++il) { + for (int alpha = 0; alpha < Nrap_read[il]; ++alpha) AveragingState.Ix2[il][alpha] = Ix2_read[il][alpha]; + } + AveragingState.Set_Label_from_Ix2(AveragingState.Ix2); + AveragingState.Compute_All(true); + + // Perform the scan: + Scan_Heis (whichDSF, AveragingState, defaultScanStatename, iKmin, iKmax, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + } + else if (Delta > 1.0) { + XXZ_gpd_Bethe_State AveragingState (chain, base); + for (int il = 0; il < chain.Nstrings; ++il) { + for (int alpha = 0; alpha < Nrap_read[il]; ++alpha) AveragingState.Ix2[il][alpha] = Ix2_read[il][alpha]; + } + AveragingState.Set_Label_from_Ix2(AveragingState.Ix2); + AveragingState.Compute_All(true); + + // Perform the scan: + Scan_Heis (whichDSF, AveragingState, defaultScanStatename, iKmin, iKmax, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + } + } + + return(0); +} + diff --git a/src/EXECS/Heis_DSF_par.cc b/src/EXECS/Heis_DSF_par.cc new file mode 100644 index 0000000..fc82d9d --- /dev/null +++ b/src/EXECS/Heis_DSF_par.cc @@ -0,0 +1,120 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Heis_DSF_par.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +#include "mpi.h" + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP Delta; + int N, M, iKneeded, iKmin, iKmax, Max_Secs; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine; + + if (argc != 8) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Heis_DSF_par executable: " << endl; + cout << endl << "This function runs ABACUS++ in parallel mode, starting from a preexisting serial run (obtained using the Heis_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: m for S- S+, z for Sz Sz, p for S+ S-." << endl; + cout << "DP Delta \t\t Value of the anisotropy: use positive real values only" << endl; + cout << "int N \t\t\t Length (number of sites) of the system: use positive even integer values only" << endl; + cout << "int M \t\t\t Number of down spins: use positive integer values between 1 and N/2" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: 0 and N" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "mpiexec -np 8 Heis_DSF_par z 1.0 100 40 0 100 600" << endl << endl; + + return(0); + } + + else { // (argc == 8) correct nr of arguments + whichDSF = *argv[1]; + Delta = atof(argv[2]); + N = atoi(argv[3]); + M = atoi(argv[4]); + iKmin = atoi(argv[5]); + iKmax = atoi(argv[6]); + Max_Secs = atoi(argv[7]); + } + + DP supercycle_time = 600.0; // allotted time per supercycle + + if (Max_Secs <= supercycle_time + 300) JSCerror("Please allow more time in Heis_DSF_par."); + + MPI::Init(argc, argv); + + DP tstart = MPI::Wtime(); + + int rank = MPI::COMM_WORLD.Get_rank(); + int nr_processors = MPI::COMM_WORLD.Get_size(); + + if (nr_processors < 2) JSCerror("Give at least 2 processors to ABACUS++ parallel !"); + + refine = true; + + // ASSUMPTION: preexisting files (raw, thr, ...) exist for the run. + + // IMPORTANT PRECONDITION: no flags are being raised in General_Scan in parallel mode, so + // the preinitializing serial run must be extensive enough to have flagged all base/type s necessary. + + DP tnow = MPI::Wtime(); + + while (tnow - tstart < Max_Secs - supercycle_time - 300) { // space for one more supercycle, + 5 minutes safety + + //cout << "rank " << rank << " ready to prepare." << endl; + + if (rank == 0) + // Split up thread list into chunks, one per processor + Prepare_Parallel_Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, nr_processors); + + //cout << "rank " << rank << " done preparing, ready to scan." << endl; + + // Barrier synchronization, to make sure other processes wait for process of rank 0 + // to have finished splitting up the thr file into pieces before starting: + MPI_Barrier (MPI::COMM_WORLD); + + // then everybody gets going on their own chunk ! + Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, + supercycle_time, target_sumrule, refine, rank, nr_processors); + + //cout << "rank " << rank << " finished scanning, reached wrapup stage." << endl; + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + // Now that everybody is done, digest data into unique files + if (rank == 0) + Wrapup_Parallel_Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, nr_processors); + + //cout << "rank " << rank << " passed wrapup stage." << endl; + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + tnow = MPI::Wtime(); + + } // while (tnow - tstart... + + + MPI::Finalize(); + + return(0); +} + diff --git a/src/EXECS/Heis_DSF_par_Prepare.cc b/src/EXECS/Heis_DSF_par_Prepare.cc new file mode 100644 index 0000000..379d540 --- /dev/null +++ b/src/EXECS/Heis_DSF_par_Prepare.cc @@ -0,0 +1,75 @@ +/********************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Heis_DSF_par_Prepare.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +//#include "mpi.h" // not needed for Prepare + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP Delta; + int N, M, iKmin, iKmax, paralevel, nr_processors_at_newlevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + if (argc < 9) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Heis_DSF_par_Prepare executable: " << endl; + cout << endl << "This function prepares for ABACUS++G in parallel mode, starting from a preexisting serial run (obtained using the Heis_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: m for S- S+, z for Sz Sz, p for S+ S-." << endl; + cout << "DP Delta \t\t Value of the anisotropy: use positive real values only" << endl; + cout << "int N \t\t\t Length (number of sites) of the system: use positive even integer values only" << endl; + cout << "int M \t\t\t Number of down spins: use positive integer values between 1 and N/2" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: 0 and N" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int nr_processors_at_new_level \t for this new parallelization level." << endl; + + return(0); + } + + else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + Delta = atof(argv[n++]); + N = atoi(argv[n++]); + M = atoi(argv[n++]); + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 9 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in Heis_DSF_par_Prepare."); + + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + nr_processors_at_newlevel = atoi(argv[n++]); + + string defaultScanStatename = ""; + + // Split up thread list into chunks, one per processor + Prepare_Parallel_Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, paralevel, rank_lower_paralevels, nr_processors_lower_paralevels, nr_processors_at_newlevel); + + } + + return(0); +} + diff --git a/src/EXECS/Heis_DSF_par_Run.cc b/src/EXECS/Heis_DSF_par_Run.cc new file mode 100644 index 0000000..7c9fc53 --- /dev/null +++ b/src/EXECS/Heis_DSF_par_Run.cc @@ -0,0 +1,127 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Heis_DSF_par_Run.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +#include "mpi.h" + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP Delta; + int N, M, iKmin, iKmax, Max_Secs, supercycle_time, paralevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + if (argc < 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Heis_DSF_par_Run executable: " << endl; + cout << endl << "This function runs ABACUS++ in parallel mode, starting from a preexisting serial run (obtained using the Heis_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: m for S- S+, z for Sz Sz, p for S+ S-." << endl; + cout << "DP Delta \t\t Value of the anisotropy: use positive real values only" << endl; + cout << "int N \t\t\t Length (number of sites) of the system: use positive even integer values only" << endl; + cout << "int M \t\t\t Number of down spins: use positive integer values between 1 and N/2" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: 0 and N" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "int supercycle_time \t\t time for one supercycle (in seconds)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "mpiexec -np 8 Heis_DSF_par_Run z 1 128 64 0 128 [**UPDATE]" << endl << endl; + + return(0); + } + + //else { // (argc == 9) correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + Delta = atof(argv[n++]); + N = atoi(argv[n++]); + M = atoi(argv[n++]); + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 10 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in Heis_DSF_par_Run."); + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + Max_Secs = atoi(argv[n++]); + supercycle_time = atoi(argv[n++]); + //} + + //DP supercycle_time = 600.0; // allotted time per supercycle + + if (Max_Secs <= supercycle_time) JSCerror("Please allow more time in Heis_DSF_par_Run."); + + MPI::Init(argc, argv); + + DP tstart = MPI::Wtime(); + + int rank_here = MPI::COMM_WORLD.Get_rank(); + int nr_processors_here = MPI::COMM_WORLD.Get_size(); + + Vect rank (paralevel); + Vect nr_processors (paralevel); + for (int i = 0; i < paralevel - 1; ++i) { + rank[i] = rank_lower_paralevels[i]; + nr_processors[i] = nr_processors_lower_paralevels[i]; + } + rank[paralevel-1] = rank_here; + nr_processors[paralevel-1] = nr_processors_here; + + if (nr_processors_here < 2) JSCerror("Give at least 2 processors to ABACUS++ parallel !"); + + refine = true; + + // ASSUMPTION: preexisting files (raw, thr, ...) exist for the run. + + + DP tnow = MPI::Wtime(); + + string defaultScanStatename = ""; + + while (tnow - tstart < Max_Secs - supercycle_time - 120) { // space for one more supercycle, + 2 minutes safety + + // Barrier synchronization, to make sure other processes wait for process of rank 0 + // to have finished splitting up the thr file into pieces before starting: + MPI_Barrier (MPI::COMM_WORLD); + + // then everybody gets going on their own chunk ! + Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, + supercycle_time, target_sumrule, refine, paralevel, rank, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + // Now that everybody is done, digest data into unique files + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + tnow = MPI::Wtime(); + + } // while (tnow - tstart... + + MPI::Finalize(); + + return(0); +} + diff --git a/src/EXECS/Heis_DSF_par_Wrapup.cc b/src/EXECS/Heis_DSF_par_Wrapup.cc new file mode 100644 index 0000000..d8013cc --- /dev/null +++ b/src/EXECS/Heis_DSF_par_Wrapup.cc @@ -0,0 +1,75 @@ +/********************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Heis_DSF_par_Prepare.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +//#include "mpi.h" // not needed for Prepare or Wrapup + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP Delta; + int N, M, iKmin, iKmax, paralevel, nr_processors_at_newlevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + if (argc < 9) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Heis_DSF_par_Wrapup executable: " << endl; + cout << endl << "This function wraps up an ABACUS++G parallel mode run." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: m for S- S+, z for Sz Sz, p for S+ S-." << endl; + cout << "DP Delta \t\t Value of the anisotropy: use positive real values only" << endl; + cout << "int N \t\t\t Length (number of sites) of the system: use positive even integer values only" << endl; + cout << "int M \t\t\t Number of down spins: use positive integer values between 1 and N/2" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: 0 and N" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int nr_processors_at_new_level \t for this latest parallelization level." << endl; + + return(0); + } + + else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + Delta = atof(argv[n++]); + N = atoi(argv[n++]); + M = atoi(argv[n++]); + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 9 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in Heis_DSF_par_Wrapup."); + + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + nr_processors_at_newlevel = atoi(argv[n++]); + + string defaultScanStatename = ""; + + // Split up thread list into chunks, one per processor + Wrapup_Parallel_Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, paralevel, rank_lower_paralevels, nr_processors_lower_paralevels, nr_processors_at_newlevel); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_Catalogue_Fixed_c_k_Nscaling.cc b/src/EXECS/LiebLin_Catalogue_Fixed_c_k_Nscaling.cc new file mode 100644 index 0000000..7295038 --- /dev/null +++ b/src/EXECS/LiebLin_Catalogue_Fixed_c_k_Nscaling.cc @@ -0,0 +1,116 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Catalogue_Fixed_c_k_Nscaling.cc + +Purpose: Produces sets of data files for correlations, increasing system size at fixed c and momentum. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 7) { // provide some info + + cout << endl << "Welcome to ABACUS\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_Catalogue_Fixed_c_k_Nscaling executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + //cout << "int Nc \t\t number of steps in interaction value" << endl; + //cout << "int Nstep \t\t\t Steps to be taken in number of particles: use positive integer values only. Filling will be set to 1 (L == N)" << endl; + //cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over (in units of N == Nstep): recommended values: 0 and 2*N" << endl; + cout << "int kfact \t\t momentum factor: momemntum will be set to kfact * kF/4" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "int Max_Secs \t\t Allowed computational time" << endl; + } + + else { // correct nr of arguments + int ia = 1; + char whichDSF = *argv[ia++]; + DP c_int = atof(argv[ia++]); + int kfact = atoi(argv[ia++]); + DP kBT = atof(argv[ia++]); + DP target_sumrule = atof(argv[ia++]); + int Max_Secs = atoi(argv[ia++]); + + + //clock_t StartTime = clock(); + double StartTime = omp_get_wtime(); + + //clock_t ActualTime = StartTime; + double ActualTime = omp_get_wtime(); + + int Secs_left = Max_Secs; + + int iN = 0; + + int nN = 12; + Vect Nv(nN); + Nv[0] = 160; Nv[1] = 192; Nv[2] = 224; Nv[3] = 256; + Nv[4] = 320; Nv[5] = 384; Nv[6] = 448; Nv[7] = 512; + Nv[8] = 640; Nv[9] = 768; Nv[10] = 896; Nv[11] = 1024; + + for (int iN = 0; iN < nN; ++iN) { + + int N = Nv[iN]; + DP L = N; + int iKmin = (kfact * N)/8; + int iKmax = iKmin; + DP srsat = 0.0; + bool refine = false; + + stringstream SRC_stringstream; string SRC_string; + Data_File_Name (SRC_stringstream, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + SRC_stringstream << ".src"; + SRC_string = SRC_stringstream.str(); const char* SRC_Cstr = SRC_string.c_str(); + + fstream srcfile; + srcfile.open(SRC_Cstr, fstream::in); + if (srcfile.fail()) { + srsat = 0.0; + refine = false; + } + else { + srcfile >> srsat; + refine = true; + } + srcfile.close(); + + ActualTime = omp_get_wtime(); + Secs_left = int(Max_Secs - (ActualTime - StartTime)); + + if (srsat < target_sumrule && Secs_left > Max_Secs/2) + // Improve the icmin calculation by one chunk: + Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, Secs_left, target_sumrule, refine); + + ActualTime = omp_get_wtime(); + + //Secs_left = int(60*Max_minutes - double(ActualTime - StartTime)/CLOCKS_PER_SEC); + Secs_left = int(Max_Secs - (ActualTime - StartTime)); + cout << "Done with N = " << N << ". Time left = " << Secs_left << " seconds." << endl; + + if (Secs_left < 60) { + cout << "Breaking out after N = " << N << " since time left = " << Secs_left << endl; + break; + } + + } // while there is time + + } // else if arguments given OK + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF.cc b/src/EXECS/LiebLin_DSF.cc new file mode 100644 index 0000000..a6f1ddd --- /dev/null +++ b/src/EXECS/LiebLin_DSF.cc @@ -0,0 +1,62 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF.cc + +Purpose: main function for ABACUS++ for LiebLin gas + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 11) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_Tgt0 executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "bool refine \t\t Is this a refinement of earlier calculations ? (0 == false, 1 == true)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "LiebLin_DSF d 1.0 100.0 100 0 200 0.56 600 1.0 0" << endl << endl; + } + + else { // (argc == 10), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + int Max_Secs = atoi(argv[8]); + DP target_sumrule = atof(argv[9]); + bool refine = (atoi(argv[10]) == 1); + + //cout << "target_sumrule = " << target_sumrule << endl; + + //Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, Max_Secs, target_sumrule, refine, 0, 1); + Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, Max_Secs, target_sumrule, refine); + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_GeneralState.cc b/src/EXECS/LiebLin_DSF_GeneralState.cc new file mode 100644 index 0000000..037abdc --- /dev/null +++ b/src/EXECS/LiebLin_DSF_GeneralState.cc @@ -0,0 +1,97 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_GeneralState.cc + +Purpose: function for ABACUS++ for LiebLin gas, on general states + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 11) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_Tgt0 executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "char* defaultScanStatename:\t\t file [].Ix2 contains the quantum numbers defining the AveragingState; used as defaultScanStatename" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + //cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "bool refine \t\t Is this a refinement of earlier calculations ? (0 == false, 1 == true)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "LiebLin_DSF d 1.0 100.0 100 0 200 0.56 600 1.0 0" << endl << endl; + } + + else { // (argc == 10), correct nr of arguments + int n = 1; + char whichDSF = *argv[n++]; + DP c_int = atof(argv[n++]); + DP L = atof(argv[n++]); + int N = atoi(argv[n++]); + char* Ix2filenameprefix = argv[n++]; + int iKmin = atoi(argv[n++]); + int iKmax = atoi(argv[n++]); + //DP kBT = atof(argv[n++]); + int Max_Secs = atoi(argv[n++]); + DP target_sumrule = atof(argv[n++]); + bool refine = (atoi(argv[n++]) == 1); + + // Read the Ix2 from the file: + Vect Ix2_input(N); + ifstream Ix2_input_file; + stringstream filenamestrstream; + filenamestrstream << Ix2filenameprefix; + string defaultScanStatename = filenamestrstream.str(); + filenamestrstream << ".Ix2"; + string filenamestr = filenamestrstream.str(); + const char* filename_Cstr = filenamestr.c_str(); + Ix2_input_file.open(filename_Cstr); + if (Ix2_input_file.fail()) { + cout << filename_Cstr << endl; + JSCerror("Could not open Ix2 input file in LiebLin_DSF_GeneralState"); + } + for (int i = 0; i < N; ++i) { + Ix2_input_file >> Ix2_input[i]; + //cout << i << "\t" << Ix2_input[i] << endl; + } + + // Now define the AveragingState + LiebLin_Bethe_State AveragingState(c_int, L, N); + AveragingState.Ix2 = Ix2_input; + AveragingState.Compute_All(true); + + //cout << "Averaging state: " << AveragingState << endl; + + //Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, Max_Secs, target_sumrule, refine, 0, 1); + //Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, Max_Secs, target_sumrule, refine); + + //void Scan_LiebLin (char whichDSF, LiebLin_Bethe_State AveragingState, string defaultScanStatename, int iKmin, int iKmax, + // int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors) + // Simplified function call of the above: + //void Scan_LiebLin (char whichDSF, LiebLin_Bethe_State AveragingState, string defaultScanStatename, int iKmin, int iKmax, + // int Max_Secs, DP target_sumrule, bool refine) + Scan_LiebLin (whichDSF, AveragingState, defaultScanStatename, iKmin, iKmax, Max_Secs, target_sumrule, refine); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_GeneralState_par_Prepare.cc b/src/EXECS/LiebLin_DSF_GeneralState_par_Prepare.cc new file mode 100644 index 0000000..6f79294 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_GeneralState_par_Prepare.cc @@ -0,0 +1,106 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_GeneralState_par_Prepare.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +//#include "mpi.h" // not needed for Prepare + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L, kBT; + int N, iKmin, iKmax, paralevel, nr_processors_at_newlevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + char* Ix2filenameprefix; + + if (argc < 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_par_Prepare executable: " << endl; + cout << endl << "This function prepares an ABACUS++ parallel mode run, starting from a preexisting serial run (obtained using the LiebLin_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "char* defaultScanStatename:\t\t file [].Ix2 contains the quantum numbers defining the AveragingState; used as defaultScanStatename" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + //cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int nr_processors_at_new_level \t for this new parallelization level." << endl; + + return(0); + } + + else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + c_int = atof(argv[n++]); + L = atof(argv[n++]); + N = atoi(argv[n++]); + Ix2filenameprefix = argv[n++]; + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + //kBT = atof(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 10 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in LiebLin_DSF_GeneralState_par_Prepare."); + + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + nr_processors_at_newlevel = atoi(argv[n++]); + + + // Read the Ix2 from the file: + Vect Ix2_input(N); + ifstream Ix2_input_file; + stringstream filenamestrstream; + filenamestrstream << Ix2filenameprefix; + string defaultScanStatename = filenamestrstream.str(); + /* + filenamestrstream << ".Ix2"; + string filenamestr = filenamestrstream.str(); + const char* filename_Cstr = filenamestr.c_str(); + Ix2_input_file.open(filename_Cstr); + if (Ix2_input_file.fail()) { + cout << filename_Cstr << endl; + JSCerror("Could not open Ix2 input file in LiebLin_DSF_GeneralState"); + } + for (int i = 0; i < N; ++i) { + Ix2_input_file >> Ix2_input[i]; + //cout << i << "\t" << Ix2_input[i] << endl; + } + + // Define the AveragingState + LiebLin_Bethe_State AveragingState(c_int, L, N); + AveragingState.Ix2 = Ix2_input; + //AveragingState.Compute_All(true); + */ + + // Split up thread list into chunks, one per processor + Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, paralevel, rank_lower_paralevels, nr_processors_lower_paralevels, nr_processors_at_newlevel); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_GeneralState_par_Run.cc b/src/EXECS/LiebLin_DSF_GeneralState_par_Run.cc new file mode 100644 index 0000000..1d19c40 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_GeneralState_par_Run.cc @@ -0,0 +1,153 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_GeneralState_par_Run.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +#include "mpi.h" + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L, kBT; + int N, iKmin, iKmax, Max_Secs, paralevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + char* Ix2filenameprefix; + + if (argc < 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_par executable: " << endl; + cout << endl << "This function runs ABACUS++ in parallel mode, starting from a preexisting serial run (obtained using the LiebLin_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "char* defaultScanStatename:\t\t file [].Ix2 contains the quantum numbers defining the AveragingState; used as defaultScanStatename" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + //cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + //cout << "int supercycle_time \t\t time for one supercycle (in seconds)" << endl; + + return(0); + } + + //else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + c_int = atof(argv[n++]); + L = atof(argv[n++]); + N = atoi(argv[n++]); + Ix2filenameprefix = argv[n++]; + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + //kBT = atof(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 10 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in LiebLin_DSF_par_Prepare."); + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + Max_Secs = atoi(argv[n++]); + // supercycle_time = atoi(argv[n++]); + //} + + //DP supercycle_time = 600.0; // allotted time per supercycle + + if (Max_Secs <= 120) JSCerror("Please allow more time in LiebLin_DSF_par_Run."); + + int Max_Secs_used = Max_Secs - 120; + + MPI::Init(argc, argv); + + DP tstart = MPI::Wtime(); + + int rank_here = MPI::COMM_WORLD.Get_rank(); + int nr_processors_here = MPI::COMM_WORLD.Get_size(); + + // Read the Ix2 from the file: + Vect Ix2_input(N); + ifstream Ix2_input_file; + stringstream filenamestrstream; + filenamestrstream << Ix2filenameprefix; + string defaultScanStatename = filenamestrstream.str(); + filenamestrstream << ".Ix2"; + string filenamestr = filenamestrstream.str(); + const char* filename_Cstr = filenamestr.c_str(); + Ix2_input_file.open(filename_Cstr); + if (Ix2_input_file.fail()) { + cout << filename_Cstr << endl; + JSCerror("Could not open Ix2 input file in LiebLin_DSF_GeneralState"); + } + for (int i = 0; i < N; ++i) { + Ix2_input_file >> Ix2_input[i]; + //cout << i << "\t" << Ix2_input[i] << endl; + } + + // Now define the AveragingState + LiebLin_Bethe_State AveragingState(c_int, L, N); + AveragingState.Ix2 = Ix2_input; + AveragingState.Compute_All(true); + + + Vect rank (paralevel); + Vect nr_processors (paralevel); + for (int i = 0; i < paralevel - 1; ++i) { + rank[i] = rank_lower_paralevels[i]; + nr_processors[i] = nr_processors_lower_paralevels[i]; + } + rank[paralevel-1] = rank_here; + nr_processors[paralevel-1] = nr_processors_here; + + if (nr_processors_here < 2) JSCerror("Give at least 2 processors to ABACUS++ parallel !"); + + refine = true; + + // ASSUMPTION: preexisting files (raw, thr, ...) exist for the run. + + + DP tnow = MPI::Wtime(); + + + //while (tnow - tstart < Max_Secs - supercycle_time - 120) { // space for one more supercycle, + 2 minutes safety + if (Max_Secs_used > 0) { + + // Barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + // then everybody gets going on their own chunk ! + //Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, + //Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, + // supercycle_time, target_sumrule, refine, paralevel, rank, nr_processors); + Scan_LiebLin (whichDSF, AveragingState, defaultScanStatename, iKmin, iKmax, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + tnow = MPI::Wtime(); + + } // while (tnow - tstart... + + MPI::Finalize(); + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_GeneralState_par_Wrapup.cc b/src/EXECS/LiebLin_DSF_GeneralState_par_Wrapup.cc new file mode 100644 index 0000000..fd361c4 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_GeneralState_par_Wrapup.cc @@ -0,0 +1,104 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_par_Prepare.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +//#include "mpi.h" // not needed for Prepare + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L, kBT; + int N, iKmin, iKmax, paralevel, nr_processors_at_newlevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + char* Ix2filenameprefix; + + if (argc < 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_par_Wrapup executable: " << endl; + cout << endl << "This function wraps up an ABACUS++ parallel mode run." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "char* defaultScanStatename:\t\t file [].Ix2 contains the quantum numbers defining the AveragingState; used as defaultScanStatename" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + //cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int nr_processors_at_new_level \t for this latest parallelization level." << endl; + + return(0); + } + + else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + c_int = atof(argv[n++]); + L = atof(argv[n++]); + N = atoi(argv[n++]); + Ix2filenameprefix = argv[n++]; + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + //kBT = atof(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 10 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in LiebLin_DSF_par_Prepare."); + + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + nr_processors_at_newlevel = atoi(argv[n++]); + + // Read the Ix2 from the file: + Vect Ix2_input(N); + ifstream Ix2_input_file; + stringstream filenamestrstream; + filenamestrstream << Ix2filenameprefix; + string defaultScanStatename = filenamestrstream.str(); + /* + filenamestrstream << ".Ix2"; + string filenamestr = filenamestrstream.str(); + const char* filename_Cstr = filenamestr.c_str(); + Ix2_input_file.open(filename_Cstr); + if (Ix2_input_file.fail()) { + cout << filename_Cstr << endl; + JSCerror("Could not open Ix2 input file in LiebLin_DSF_GeneralState"); + } + for (int i = 0; i < N; ++i) { + Ix2_input_file >> Ix2_input[i]; + //cout << i << "\t" << Ix2_input[i] << endl; + } + + // Define the AveragingState + LiebLin_Bethe_State AveragingState(c_int, L, N); + AveragingState.Ix2 = Ix2_input; + //AveragingState.Compute_All(true); + */ + + // Digest files into a unique one for the latest paralevel: + Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, paralevel, rank_lower_paralevels, nr_processors_lower_paralevels, nr_processors_at_newlevel); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_MosesState.cc b/src/EXECS/LiebLin_DSF_MosesState.cc new file mode 100644 index 0000000..dd791e8 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_MosesState.cc @@ -0,0 +1,84 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF.cc + +Purpose: main function for ABACUS++ for LiebLin gas + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 13) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_MosesState executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int Nl \t\t\t Number of particles in left Fermi sea (Nr is then N - Nl)" << endl; + cout << "int DIl \t\t shift of left sea as compared to its ground state position" << endl; + cout << "int DIr \t\t shift of right sea as compared to its ground state position" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + //cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "bool refine \t\t Is this a refinement of earlier calculations ? (0 == false, 1 == true)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "LiebLin_DSF_MosesState d 1.0 100.0 100 50 -30 20 0 200 600 1.0 0" << endl << endl; + } + + else { // (argc == 13), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int Nl = atoi(argv[5]); + //int Nr = N - Nl; + int DIl = atoi(argv[6]); + int DIr = atoi(argv[7]); + int iKmin = atoi(argv[8]); + int iKmax = atoi(argv[9]); + //DP kBT = atof(argv[7]); + int Max_Secs = atoi(argv[10]); + DP target_sumrule = atof(argv[11]); + bool refine = (atoi(argv[12]) == 1); + + // Define the Moses state: + LiebLin_Bethe_State MosesState (c_int, L, N); + + // Split the sea: + for (int i = 0; i < Nl; ++i) MosesState.Ix2[i] += 2 * DIl; + for (int i = Nl; i < N; ++i) MosesState.Ix2[i] += 2 * DIr; + + MosesState.Compute_All (true); + + //cout << MosesState << endl; + + // Handy default name: + stringstream defaultScanStatename_strstream; + defaultScanStatename_strstream << "Moses_Nl_" << Nl << "_DIl_" << DIl << "_DIr_" << DIr; + string defaultScanStatename = defaultScanStatename_strstream.str(); + + // Compute the correlation: + Scan_LiebLin (whichDSF, MosesState, defaultScanStatename, iKmin, iKmax, Max_Secs, target_sumrule, refine); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_MosesState_par.cc b/src/EXECS/LiebLin_DSF_MosesState_par.cc new file mode 100644 index 0000000..3553e40 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_MosesState_par.cc @@ -0,0 +1,139 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_par.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +#include "mpi.h" + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L; + int N, Nl, DIl, DIr, iKmin, iKmax, Max_Secs, supercycle_time; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + DP kBT = 0.0; // dummy + + if (argc != 12) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_MosesState_par executable: " << endl; + cout << endl << "This function runs ABACUS++ in parallel mode, starting from a preexisting serial run (obtained using the LiebLin_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int Nl \t\t\t Number of particles in left Fermi sea (Nr is then N - Nl)" << endl; + cout << "int DIl \t\t shift of left sea as compared to its ground state position" << endl; + cout << "int DIr \t\t shift of right sea as compared to its ground state position" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "int supercycle_time \t\t time for one supercycle (in seconds)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "mpiexec -np 8 LiebLin_DSF_MosesState_par d 1.0 100.0 100 50 -30 20 -400 400 3600 600" << endl << endl; + + return(0); + } + + else { // (argc == 11) correct nr of arguments + whichDSF = *argv[1]; + c_int = atof(argv[2]); + L = atof(argv[3]); + N = atoi(argv[4]); + Nl = atoi(argv[5]); + DIl = atoi(argv[6]); + DIr = atoi(argv[7]); + iKmin = atoi(argv[8]); + iKmax = atoi(argv[9]); + Max_Secs = atoi(argv[10]); + supercycle_time = atoi(argv[11]); + } + + //DP supercycle_time = 600.0; // allotted time per supercycle + + if (Max_Secs <= supercycle_time) JSCerror("Please allow more time in LiebLin_DSF_par."); + + MPI::Init(argc, argv); + + DP tstart = MPI::Wtime(); + + int rank = MPI::COMM_WORLD.Get_rank(); + int nr_processors = MPI::COMM_WORLD.Get_size(); + + if (nr_processors < 2) JSCerror("Give at least 2 processors to ABACUS++ parallel !"); + + refine = true; + + // ASSUMPTION: preexisting files (raw, thr, ...) exist for the run. + + + DP tnow = MPI::Wtime(); + + // Define the Moses state: + LiebLin_Bethe_State MosesState (c_int, L, N); + + // Split the sea: + for (int i = 0; i < Nl; ++i) MosesState.Ix2[i] += 2 * DIl; + for (int i = Nl; i < N; ++i) MosesState.Ix2[i] += 2 * DIr; + + MosesState.Compute_All (true); + + // Handy default name: + stringstream defaultScanStatename_strstream; + defaultScanStatename_strstream << "Moses_Nl_" << Nl << "_DIl_" << DIl << "_DIr_" << DIr; + string defaultScanStatename = defaultScanStatename_strstream.str(); + + MPI_Barrier (MPI::COMM_WORLD); + + while (tnow - tstart < Max_Secs - supercycle_time - 120) { // space for one more supercycle, + 2 minutes safety + + if (rank == 0) + // Split up thread list into chunks, one per processor + //Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, nr_processors); + Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, nr_processors); + + // Barrier synchronization, to make sure other processes wait for process of rank 0 + // to have finished splitting up the thr file into pieces before starting: + MPI_Barrier (MPI::COMM_WORLD); + + // then everybody gets going on their own chunk ! + //Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, + //Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, supercycle_time, target_sumrule, refine, rank, nr_processors); + Scan_LiebLin (whichDSF, MosesState, defaultScanStatename, iKmin, iKmax, supercycle_time, target_sumrule, refine, rank, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + // Now that everybody is done, digest data into unique files + + if (rank == 0) + //Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, nr_processors); + Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + tnow = MPI::Wtime(); + + } // while (tnow - tstart... + + MPI::Finalize(); + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_MosesState_par_Prepare.cc b/src/EXECS/LiebLin_DSF_MosesState_par_Prepare.cc new file mode 100644 index 0000000..2eea6e1 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_MosesState_par_Prepare.cc @@ -0,0 +1,96 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_par_Prepare.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +//#include "mpi.h" // not needed for Prepare + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L; + int N, Nl, DIl, DIr, iKmin, iKmax, paralevel, nr_processors_at_newlevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + DP kBT = 0.0; // dummy + + if (argc < 12) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_MosesState_par_Prepare executable: " << endl; + cout << endl << "This function prepares an ABACUS++ parallel mode run, starting from a preexisting serial run (obtained using the LiebLin_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int Nl \t\t\t Number of particles in left Fermi sea (Nr is then N - Nl)" << endl; + cout << "int DIl \t\t shift of left sea as compared to its ground state position" << endl; + cout << "int DIr \t\t shift of right sea as compared to its ground state position" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int nr_processors_at_new_level \t for this new parallelization level." << endl; + + return(0); + } + + else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + c_int = atof(argv[n++]); + L = atof(argv[n++]); + N = atoi(argv[n++]); + Nl = atoi(argv[n++]); + DIl = atoi(argv[n++]); + DIr = atoi(argv[n++]); + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 12 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in LiebLin_DSF_MosesState_par_Prepare."); + + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + nr_processors_at_newlevel = atoi(argv[n++]); + + // Define the Moses state: + LiebLin_Bethe_State MosesState (c_int, L, N); + + // Split the sea: + for (int i = 0; i < Nl; ++i) MosesState.Ix2[i] += 2 * DIl; + for (int i = Nl; i < N; ++i) MosesState.Ix2[i] += 2 * DIr; + + MosesState.Compute_All (true); + + // Handy default name: + stringstream defaultScanStatename_strstream; + defaultScanStatename_strstream << "Moses_Nl_" << Nl << "_DIl_" << DIl << "_DIr_" << DIr; + string defaultScanStatename = defaultScanStatename_strstream.str(); + + + // Split up thread list into chunks, one per processor + Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, paralevel, rank_lower_paralevels, nr_processors_lower_paralevels, nr_processors_at_newlevel); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_MosesState_par_Run.cc b/src/EXECS/LiebLin_DSF_MosesState_par_Run.cc new file mode 100644 index 0000000..c63de9b --- /dev/null +++ b/src/EXECS/LiebLin_DSF_MosesState_par_Run.cc @@ -0,0 +1,162 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_par.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +#include "mpi.h" + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L; + int N, Nl, DIl, DIr, iKmin, iKmax, Max_Secs, supercycle_time, paralevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + DP kBT = 0.0; // dummy + + if (argc < 13) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_MosesState_par_Run executable: " << endl; + cout << endl << "This function runs ABACUS++ in parallel mode, starting from a preexisting serial run (obtained using the LiebLin_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int Nl \t\t\t Number of particles in left Fermi sea (Nr is then N - Nl)" << endl; + cout << "int DIl \t\t shift of left sea as compared to its ground state position" << endl; + cout << "int DIr \t\t shift of right sea as compared to its ground state position" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "int supercycle_time \t\t time for one supercycle (in seconds)" << endl; + + return(0); + } + + //else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + c_int = atof(argv[n++]); + L = atof(argv[n++]); + N = atoi(argv[n++]); + Nl = atoi(argv[n++]); + DIl = atoi(argv[n++]); + DIr = atoi(argv[n++]); + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 13 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in LiebLin_DSF_par_Prepare."); + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + Max_Secs = atoi(argv[n++]); + supercycle_time = atoi(argv[n++]); + //} + + //DP supercycle_time = 600.0; // allotted time per supercycle + + if (Max_Secs <= supercycle_time) JSCerror("Please allow more time in LiebLin_DSF_par_Run."); + + MPI::Init(argc, argv); + + DP tstart = MPI::Wtime(); + + int rank_here = MPI::COMM_WORLD.Get_rank(); + int nr_processors_here = MPI::COMM_WORLD.Get_size(); + + //cout << "rank " << rank_here << " out of " << nr_processors_here << " ready to go." << endl; + + Vect rank (paralevel); + Vect nr_processors (paralevel); + for (int i = 0; i < paralevel - 1; ++i) { + rank[i] = rank_lower_paralevels[i]; + nr_processors[i] = nr_processors_lower_paralevels[i]; + } + rank[paralevel-1] = rank_here; + nr_processors[paralevel-1] = nr_processors_here; + + if (nr_processors_here < 2) JSCerror("Give at least 2 processors to ABACUS++ parallel !"); + + refine = true; + + // ASSUMPTION: preexisting files (raw, thr, ...) exist for the run. + + + DP tnow = MPI::Wtime(); + + // Define the Moses state: + LiebLin_Bethe_State MosesState (c_int, L, N); + + // Split the sea: + for (int i = 0; i < Nl; ++i) MosesState.Ix2[i] += 2 * DIl; + for (int i = Nl; i < N; ++i) MosesState.Ix2[i] += 2 * DIr; + + MosesState.Compute_All (true); + + // Handy default name: + stringstream defaultScanStatename_strstream; + defaultScanStatename_strstream << "Moses_Nl_" << Nl << "_DIl_" << DIl << "_DIr_" << DIr; + string defaultScanStatename = defaultScanStatename_strstream.str(); + + //cout << "rank " << rank_here << " out of " << nr_processors_here << " waiting at barrier." << endl; + + MPI_Barrier (MPI::COMM_WORLD); + + + while (tnow - tstart < Max_Secs - supercycle_time - 120) { // space for one more supercycle, + 2 minutes safety + + //if (rank == 0) + // Split up thread list into chunks, one per processor + //Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, nr_processors); + //Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, nr_processors); + + // Barrier synchronization, to make sure other processes wait for process of rank 0 + // to have finished splitting up the thr file into pieces before starting: + MPI_Barrier (MPI::COMM_WORLD); + + // then everybody gets going on their own chunk ! + //Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, + //Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, supercycle_time, target_sumrule, refine, paralevel, rank, nr_processors); + Scan_LiebLin (whichDSF, MosesState, defaultScanStatename, iKmin, iKmax, supercycle_time, target_sumrule, refine, paralevel, rank, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + // Now that everybody is done, digest data into unique files + + //if (rank == 0) + //Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, nr_processors); + //Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + tnow = MPI::Wtime(); + + } // while (tnow - tstart... + + MPI::Finalize(); + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_MosesState_par_Wrapup.cc b/src/EXECS/LiebLin_DSF_MosesState_par_Wrapup.cc new file mode 100644 index 0000000..f7fae54 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_MosesState_par_Wrapup.cc @@ -0,0 +1,96 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_MosesState_par_Wrapup.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +//#include "mpi.h" // not needed for Prepare + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L; + int N, Nl, DIl, DIr, iKmin, iKmax, paralevel, nr_processors_at_newlevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + DP kBT = 0.0; // dummy + + if (argc < 12) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_MosesState_par_Wrapup executable: " << endl; + cout << endl << "This function wraps up an ABACUS++ parallel mode run." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int Nl \t\t\t Number of particles in left Fermi sea (Nr is then N - Nl)" << endl; + cout << "int DIl \t\t shift of left sea as compared to its ground state position" << endl; + cout << "int DIr \t\t shift of right sea as compared to its ground state position" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int nr_processors_at_new_level \t for this new parallelization level." << endl; + + return(0); + } + + else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + c_int = atof(argv[n++]); + L = atof(argv[n++]); + N = atoi(argv[n++]); + Nl = atoi(argv[n++]); + DIl = atoi(argv[n++]); + DIr = atoi(argv[n++]); + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 12 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in LiebLin_DSF_MosesState_par_Prepare."); + + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + nr_processors_at_newlevel = atoi(argv[n++]); + + // Define the Moses state: + LiebLin_Bethe_State MosesState (c_int, L, N); + + // Split the sea: + for (int i = 0; i < Nl; ++i) MosesState.Ix2[i] += 2 * DIl; + for (int i = Nl; i < N; ++i) MosesState.Ix2[i] += 2 * DIr; + + MosesState.Compute_All (true); + + // Handy default name: + stringstream defaultScanStatename_strstream; + defaultScanStatename_strstream << "Moses_Nl_" << Nl << "_DIl_" << DIl << "_DIr_" << DIr; + string defaultScanStatename = defaultScanStatename_strstream.str(); + + + // Digest files into a unique one for the latest paralevel: + Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, paralevel, rank_lower_paralevels, nr_processors_lower_paralevels, nr_processors_at_newlevel); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_over_Ensemble.cc b/src/EXECS/LiebLin_DSF_over_Ensemble.cc new file mode 100644 index 0000000..9ba8934 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_over_Ensemble.cc @@ -0,0 +1,99 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_over_Ensemble.cc + +Purpose: main function for ABACUS++T for LiebLin gas, averaging over an Ensemble. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_Tgt0 executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + //cout << "int nstates \t\t\t Number of states to be considered in the ensemble" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + //cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "bool refine \t\t Is this a refinement of earlier calculations ? (0 == false, 1 == true)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "LiebLin_DSF_over_Ensemble d 1.0 100.0 100 0 200 0.56 10 600 0" << endl << endl; + } + + else { // (argc == 10), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + //int nstates_req = atoi(argv[8]); + int Max_Secs = atoi(argv[8]); + bool refine = (atoi(argv[9]) == 1); + + // Start by constructing (or loading) the state ensemble. + + LiebLin_Diagonal_State_Ensemble ensemble; + + stringstream ensfilestrstream; + //ensfilestrstream << "LiebLin_c_int_" << c_int << "_L_" << L << "_N_" << N << "_kBT_" << kBT << "_ns_" << nstates_req << ".ens"; + ensfilestrstream << "LiebLin_c_int_" << c_int << "_L_" << L << "_N_" << N << "_kBT_" << kBT << ".ens"; + string ensfilestr = ensfilestrstream.str(); + const char* ensfile_Cstr = ensfilestr.c_str(); + + if (!refine) { // Construct the state ensemble + //ensemble = LiebLin_Thermal_Saddle_Point_Ensemble (c_int, L, N, kBT, nstates_req); + ensemble = LiebLin_Thermal_Saddle_Point_Ensemble (c_int, L, N, kBT); + ensemble.Save(ensfile_Cstr); // Save the ensemble + } + + else { // load the ensemble data + ensemble.Load(c_int, L, N, ensfile_Cstr); + } + + // Now perform the DSF calculation over each state in the ensemble, distributing the time according to the weight + + for (int ns = 0; ns < ensemble.nstates; ++ns) { + //void Scan_LiebLin (char whichDSF, LiebLin_Bethe_State AveragingState, string defaultScanStatename, int iKmin, int iKmax, + //int Max_Secs, DP target_sumrule, bool refine, int rank, int nr_processors) + //Scan_LiebLin (whichDSF, ensemble.state[ns], ensemble.state[ns].label, iKmin, iKmax, int(Max_Secs * ensemble.weight[ns]), 1.0e+6, refine, 0, 1); + Scan_LiebLin (whichDSF, ensemble.state[ns], ensemble.state[ns].label, iKmin, iKmax, int(Max_Secs * ensemble.weight[ns]), 1.0e+6, refine); + } + + // Evaluate the f-sumrule + stringstream FSR_stringstream; string FSR_string; + Data_File_Name (FSR_stringstream, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + FSR_stringstream << "_ns_" << ensemble.nstates << ".fsr"; + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + + DP Chem_Pot = 0.0; + + Evaluate_F_Sumrule (whichDSF, c_int, L, N, kBT, ensemble.nstates, Chem_Pot, iKmin, iKmax, FSR_Cstr); + + + } // correct nr of arguments + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_over_Ensemble_par.cc b/src/EXECS/LiebLin_DSF_over_Ensemble_par.cc new file mode 100644 index 0000000..0f2b32e --- /dev/null +++ b/src/EXECS/LiebLin_DSF_over_Ensemble_par.cc @@ -0,0 +1,191 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_over_Ensemble_par.cc + +Purpose: main function for ABACUS for LiebLin gas, averaging over an Ensemble, parallel implementation. + +***********************************************************/ + +#include "JSC.h" +#include "mpi.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_Tgt0 executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + //cout << "int nstates \t\t\t Number of states to be considered in the ensemble" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + //cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "bool refine \t\t Is this a refinement of earlier calculations ? (0 == false, 1 == true)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "LiebLin_DSF_over_Ensemble d 1.0 100.0 100 0 200 0.56 10 600 0" << endl << endl; + } + + else { // (argc == 10), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + //int nstates_req = atoi(argv[8]); + int Max_Secs = atoi(argv[8]); + bool refine = (atoi(argv[9]) == 1); + + if (refine == false) JSCerror("Please run the serial version of LiebLin_DSF_over_Ensemble first."); + + MPI::Init(argc, argv); + + DP tstart = MPI::Wtime(); + + int rank = MPI::COMM_WORLD.Get_rank(); + int nr_processors = MPI::COMM_WORLD.Get_size(); + + if (nr_processors < 2) JSCerror("Give at least 2 processors to ABACUS++ parallel !"); + + + // Start by constructing (or loading) the state ensemble. + + LiebLin_Diagonal_State_Ensemble ensemble; + + stringstream ensfilestrstream; + //ensfilestrstream << "LiebLin_c_int_" << c_int << "_L_" << L << "_N_" << N << "_kBT_" << kBT << "_ns_" << nstates_req << ".ens"; + ensfilestrstream << "LiebLin_c_int_" << c_int << "_L_" << L << "_N_" << N << "_kBT_" << kBT << ".ens"; + string ensfilestr = ensfilestrstream.str(); + const char* ensfile_Cstr = ensfilestr.c_str(); + + if (!refine) { // Construct the state ensemble + //ensemble = LiebLin_Thermal_Saddle_Point_Ensemble (c_int, L, N, kBT, nstates_req); + ensemble = LiebLin_Thermal_Saddle_Point_Ensemble (c_int, L, N, kBT); + ensemble.Save(ensfile_Cstr); // Save the ensemble + } + + else { // load the ensemble data + ensemble.Load(c_int, L, N, ensfile_Cstr); + } + + MPI_Barrier (MPI::COMM_WORLD); + + // Now perform the DSF calculation over each state in the ensemble + + /* Original implementation: Scan always called serially. Superseded by version below, using successive parallel scans on each state in the ensemble. + int nDSFperproc = ensemble.nstates/nr_processors + 1; + //if (ensemble.nstates % nr_processors) JSCerror("Use nr_processors * integer multiple == ensemble.nstates in LiebLin_DSF_over_Ensemble_par."); + + // Processor with rank r does all + + int ns; + int Max_Secs_used = Max_Secs/nDSFperproc; + + for (int ir = 0; ir < nDSFperproc; ++ir) { + ns = rank + ir * nr_processors; + //void Scan_LiebLin (char whichDSF, LiebLin_Bethe_State AveragingState, string defaultScanStatename, int iKmin, int iKmax, + //int Max_Secs, DP target_sumrule, bool refine, int rank, int nr_processors) + if (ns < ensemble.nstates) { + //cout << "Processor rank " << rank << " going for ns = " << ns << " out of " << ensemble.nstates << endl; + Scan_LiebLin (whichDSF, ensemble.state[ns], ensemble.state[ns].label, iKmin, iKmax, Max_Secs_used, 1.0e+6, refine, 0, 1); + } + } + */ + + // Version 2013 04 24: + // Makes use of a parallel scan for each state in the ensemble, in succession. + // Code is simple adaptation of LiebLin_DSF_par executable code. + + int Max_Secs_used = Max_Secs/ensemble.nstates; + + DP supercycle_time = 600.0; // allotted time per supercycle + + if (Max_Secs_used <= supercycle_time) JSCerror("Please allow more time in LiebLin_DSF_par."); + + // Main loop over ensemble: + for (int ns = 0; ns < ensemble.nstates; ++ns) { + + tstart = MPI::Wtime(); + DP tnow = MPI::Wtime(); + + string defaultScanStatename = ensemble.state[ns].label; + + while (tnow - tstart < Max_Secs_used - supercycle_time) { // space for one more supercycle + + if (rank == 0) + // Split up thread list into chunks, one per processor + //Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, nr_processors); + Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, nr_processors); + + // Barrier synchronization, to make sure other processes wait for process of rank 0 + // to have finished splitting up the thr file into pieces before starting: + MPI_Barrier (MPI::COMM_WORLD); + + // then everybody gets going on their own chunk ! + //Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, + //Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, + //supercycle_time, target_sumrule, refine, rank, nr_processors); + Scan_LiebLin (whichDSF, ensemble.state[ns], ensemble.state[ns].label, iKmin, iKmax, supercycle_time, 1.0e+6, refine, rank, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + // Now that everybody is done, digest data into unique files + + if (rank == 0) + //Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, nr_processors); + Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + tnow = MPI::Wtime(); + + } // while (tnow - tstart... + + } // for ns + + + MPI_Barrier (MPI::COMM_WORLD); + + + // Final wrapup of the data + if (rank == 0) { + + // Evaluate the f-sumrule + stringstream FSR_stringstream; string FSR_string; + Data_File_Name (FSR_stringstream, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + FSR_stringstream << "_ns_" << ensemble.nstates << ".fsr"; + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + + DP Chem_Pot = 0.0; + + Evaluate_F_Sumrule (whichDSF, c_int, L, N, kBT, ensemble.nstates, Chem_Pot, iKmin, iKmax, FSR_Cstr); + } + + MPI_Barrier (MPI::COMM_WORLD); + + } // correct nr of arguments + + MPI::Finalize(); + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_par.cc b/src/EXECS/LiebLin_DSF_par.cc new file mode 100644 index 0000000..9903e2b --- /dev/null +++ b/src/EXECS/LiebLin_DSF_par.cc @@ -0,0 +1,119 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_par.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +#include "mpi.h" + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L, kBT; + int N, iKmin, iKmax, Max_Secs, supercycle_time; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + if (argc != 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_par executable: " << endl; + cout << endl << "This function runs ABACUS++ in parallel mode, starting from a preexisting serial run (obtained using the LiebLin_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "int supercycle_time \t\t time for one supercycle (in seconds)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "mpiexec -np 8 LiebLin_DSF_MosesState_par d 1.0 100.0 100 -400 400 0.0 3600 600" << endl << endl; + + return(0); + } + + else { // (argc == 9) correct nr of arguments + whichDSF = *argv[1]; + c_int = atof(argv[2]); + L = atof(argv[3]); + N = atoi(argv[4]); + iKmin = atoi(argv[5]); + iKmax = atoi(argv[6]); + kBT = atof(argv[7]); + Max_Secs = atoi(argv[8]); + supercycle_time = atoi(argv[9]); + } + + //DP supercycle_time = 600.0; // allotted time per supercycle + + if (Max_Secs <= supercycle_time) JSCerror("Please allow more time in LiebLin_DSF_par."); + + MPI::Init(argc, argv); + + DP tstart = MPI::Wtime(); + + int rank = MPI::COMM_WORLD.Get_rank(); + int nr_processors = MPI::COMM_WORLD.Get_size(); + + if (nr_processors < 2) JSCerror("Give at least 2 processors to ABACUS++ parallel !"); + + refine = true; + + // ASSUMPTION: preexisting files (raw, thr, ...) exist for the run. + + + DP tnow = MPI::Wtime(); + + string defaultScanStatename = ""; + + while (tnow - tstart < Max_Secs - supercycle_time - 120) { // space for one more supercycle, + 2 minutes safety + + if (rank == 0) + // Split up thread list into chunks, one per processor + //Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, nr_processors); + Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, nr_processors); + + // Barrier synchronization, to make sure other processes wait for process of rank 0 + // to have finished splitting up the thr file into pieces before starting: + MPI_Barrier (MPI::COMM_WORLD); + + // then everybody gets going on their own chunk ! + //Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, + Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, + supercycle_time, target_sumrule, refine, rank, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + // Now that everybody is done, digest data into unique files + + if (rank == 0) + //Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, nr_processors); + Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + tnow = MPI::Wtime(); + + } // while (tnow - tstart... + + MPI::Finalize(); + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_par_Prepare.cc b/src/EXECS/LiebLin_DSF_par_Prepare.cc new file mode 100644 index 0000000..5fe5a21 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_par_Prepare.cc @@ -0,0 +1,77 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_par_Prepare.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +//#include "mpi.h" // not needed for Prepare + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L, kBT; + int N, iKmin, iKmax, paralevel, nr_processors_at_newlevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + if (argc < 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_par_Prepare executable: " << endl; + cout << endl << "This function prepares an ABACUS++ parallel mode run, starting from a preexisting serial run (obtained using the LiebLin_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int nr_processors_at_new_level \t for this new parallelization level." << endl; + + return(0); + } + + else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + c_int = atof(argv[n++]); + L = atof(argv[n++]); + N = atoi(argv[n++]); + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + kBT = atof(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 10 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in LiebLin_DSF_par_Prepare."); + + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + nr_processors_at_newlevel = atoi(argv[n++]); + + string defaultScanStatename = ""; + + // Split up thread list into chunks, one per processor + Prepare_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, paralevel, rank_lower_paralevels, nr_processors_lower_paralevels, nr_processors_at_newlevel); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_par_Run.cc b/src/EXECS/LiebLin_DSF_par_Run.cc new file mode 100644 index 0000000..e9a42f7 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_par_Run.cc @@ -0,0 +1,120 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_par_Run.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +#include "mpi.h" + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L, kBT; + int N, iKmin, iKmax, Max_Secs, paralevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + if (argc < 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_par_Run executable: " << endl; + cout << endl << "This function runs ABACUS++ in parallel mode, starting from a preexisting serial run (obtained using the LiebLin_DSF executable) using the same model parameters." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "mpiexec -np 8 LiebLin_DSF_par_Run d 1.0 100.0 100 0 400 0.0 1 3600" << endl << endl; + + return(0); + } + + int n = 1; + whichDSF = *argv[n++]; + c_int = atof(argv[n++]); + L = atof(argv[n++]); + N = atoi(argv[n++]); + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + kBT = atof(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 10 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in LiebLin_DSF_par_Run."); + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + Max_Secs = atoi(argv[n++]); + + if (Max_Secs < 120) JSCerror("Please allow more time in LiebLin_DSF_par_Run."); + + int Max_Secs_used = Max_Secs - 120; + + MPI::Init(argc, argv); + + DP tstart = MPI::Wtime(); + + int rank_here = MPI::COMM_WORLD.Get_rank(); + int nr_processors_here = MPI::COMM_WORLD.Get_size(); + + Vect rank (paralevel); + Vect nr_processors (paralevel); + for (int i = 0; i < paralevel - 1; ++i) { + rank[i] = rank_lower_paralevels[i]; + nr_processors[i] = nr_processors_lower_paralevels[i]; + } + rank[paralevel-1] = rank_here; + nr_processors[paralevel-1] = nr_processors_here; + + if (nr_processors_here < 2) JSCerror("Give at least 2 processors to ABACUS parallel !"); + + refine = true; + + // ASSUMPTION: preexisting files (raw, thr, ...) exist for the run. + + + DP tnow = MPI::Wtime(); + + string defaultScanStatename = ""; + + if (Max_Secs_used > 0) { // space for 2 minutes safety + + // Barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + // then everybody gets going on their own chunk ! + //Scan_LiebLin (whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, + Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, + Max_Secs_used, target_sumrule, refine, paralevel, rank, nr_processors); + + // Another barrier synchronization + MPI_Barrier (MPI::COMM_WORLD); + + tnow = MPI::Wtime(); + + } + + MPI::Finalize(); + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_par_Wrapup.cc b/src/EXECS/LiebLin_DSF_par_Wrapup.cc new file mode 100644 index 0000000..51c5d45 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_par_Wrapup.cc @@ -0,0 +1,77 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_par_Wrapup.cc + +Purpose: Parallel version of ABACUS++ using MPICH. + + +***********************************************************/ + +#include "JSC.h" +//#include "mpi.h" // not needed for Prepare or Wrapup + +using namespace JSC; + +int main(int argc, char *argv[]) +{ + char whichDSF; + DP c_int, L, kBT; + int N, iKmin, iKmax, paralevel, nr_processors_at_newlevel; + DP target_sumrule = 1.0e+6; // effectively deactivated here + bool refine = true; // always true for parallel mode + + if (argc < 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_par_Wrapup executable: " << endl; + cout << endl << "This function wraps up an ABACUS++ parallel mode run." << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int paralevel" << endl; + cout << "rank[i], nr_processors[i] \t rank and nr_processors of each earlier paralevels." << endl; + cout << "int nr_processors_at_new_level \t for this latest parallelization level." << endl; + + return(0); + } + + else { // correct nr of arguments + int n = 1; + whichDSF = *argv[n++]; + c_int = atof(argv[n++]); + L = atof(argv[n++]); + N = atoi(argv[n++]); + iKmin = atoi(argv[n++]); + iKmax = atoi(argv[n++]); + kBT = atof(argv[n++]); + paralevel = atoi(argv[n++]); // paralevel == 1 means that we have one layer of parallelization, so no previous rank and nr_processors to specify + if (argc != 10 + 2*(paralevel - 1)) JSCerror("Wrong nr of arguments in LiebLin_DSF_par_Wrapup."); + + Vect rank_lower_paralevels(paralevel - 1); + Vect nr_processors_lower_paralevels(paralevel - 1); + for (int i = 0; i < paralevel - 1; ++i) { + rank_lower_paralevels[i] = atoi(argv[n++]); + nr_processors_lower_paralevels[i] = atoi(argv[n++]); + } + nr_processors_at_newlevel = atoi(argv[n++]); + + string defaultScanStatename = ""; + + // Digest files into a unique one for the latest paralevel: + Wrapup_Parallel_Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, defaultScanStatename, paralevel, rank_lower_paralevels, nr_processors_lower_paralevels, nr_processors_at_newlevel); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_tester.cc b/src/EXECS/LiebLin_DSF_tester.cc new file mode 100644 index 0000000..a68c4f9 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_tester.cc @@ -0,0 +1,92 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_tester.cc + +Purpose: allows for Ix2 manipulations (user-prompted) for LiebLin gas + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 6) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_tester executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "LiebLin_DSF_tester d 1.0 100.0 100 0.56 " << endl << endl; + } + + else { // (argc == 6), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + DP kBT = atof(argv[5]); + + //if (whichDSF != 'd') JSCerror("Other options not implemented yet in finite T Scan_LiebLin"); + + // Delta is the number of sites involved in the smoothing of the entropy + //int Delta = int(sqrt(N))/2;//6;//N/20; + + // Construct the finite-size saddle-point state: + //LiebLin_Bethe_State spstate = Canonical_Saddle_Point_State (c_int, L, N, kBT, Delta); + LiebLin_Bethe_State spstate = Canonical_Saddle_Point_State (c_int, L, N, kBT); + spstate.Compute_All(true); + + LiebLin_Bethe_State estate = spstate; + if (whichDSF == 'o') estate = Remove_Particle_at_Center (spstate); + else if (whichDSF == 'g') estate = Add_Particle_at_Center (spstate); + if (whichDSF != 'd') estate.Compute_All(true); + cout << estate << endl; + Vect OriginIx2 = estate.Ix2; + + int Ix2old, Ix2new; + int again = 0; + string label_req; + do { + //cout << "Substitute Ix2: which for which ?" << endl; + //cin >> Ix2old >> Ix2new; + //for (int i = 0; i < estate.N; ++i) if (estate.Ix2[i] == Ix2old) estate.Ix2[i] = Ix2new; + //estate.Ix2.QuickSort(); + cout << "Which label should the exc state be set to?" << endl; + cin >> label_req; + estate.Set_to_Label(label_req, OriginIx2); + estate.Compute_All(false); + cout << spstate << endl; + cout << estate; + if (whichDSF == 'd') + cout << setprecision(16) << "estate.E - spstate.E = " << estate.E - spstate.E << "\tME = " << real(exp(ln_Density_ME(spstate, estate))) << endl; + else if (whichDSF == 'o') + cout << setprecision(16) << "estate.E - spstate.E = " << estate.E - spstate.E << "\tME = " << real(exp(ln_Psi_ME(estate, spstate))) << endl; + else if (whichDSF == 'g') + cout << setprecision(16) << "estate.E - spstate.E = " << estate.E - spstate.E << "\tME = " << real(exp(ln_Psi_ME(spstate, estate))) << endl; + + //cout << "Another try ? (1 == yes, 0 == no)" << endl; + again = 1; + cin >> again; + } while (again != 0); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_DSF_tester_Ix2.cc b/src/EXECS/LiebLin_DSF_tester_Ix2.cc new file mode 100644 index 0000000..99aada4 --- /dev/null +++ b/src/EXECS/LiebLin_DSF_tester_Ix2.cc @@ -0,0 +1,92 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_tester.cc + +Purpose: allows for Ix2 manipulations (user-prompted) for LiebLin gas + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 6) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_tester executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "LiebLin_DSF_tester d 1.0 100.0 100 0.56 " << endl << endl; + } + + else { // (argc == 6), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + DP kBT = atof(argv[5]); + + //if (whichDSF != 'd') JSCerror("Other options not implemented yet in finite T Scan_LiebLin"); + + // Delta is the number of sites involved in the smoothing of the entropy + //int Delta = int(sqrt(N))/2;//6;//N/20; + + // Construct the finite-size saddle-point state: + //LiebLin_Bethe_State spstate = Canonical_Saddle_Point_State (c_int, L, N, kBT, Delta); + LiebLin_Bethe_State spstate = Canonical_Saddle_Point_State (c_int, L, N, kBT); + spstate.Compute_All(true); + + LiebLin_Bethe_State estate = spstate; + if (whichDSF == 'o') estate = Remove_Particle_at_Center (spstate); + else if (whichDSF == 'g') estate = Add_Particle_at_Center (spstate); + if (whichDSF != 'd') estate.Compute_All(true); + cout << estate << endl; + Vect OriginIx2 = estate.Ix2; + + int Ix2old, Ix2new; + int again = 0; + string label_req; + do { + cout << "Substitute Ix2: which for which ?" << endl; + cin >> Ix2old >> Ix2new; + for (int i = 0; i < estate.N; ++i) if (estate.Ix2[i] == Ix2old) estate.Ix2[i] = Ix2new; + estate.Ix2.QuickSort(); + //cout << "Which label should the exc state be set to?" << endl; + //cin >> label_req; + //estate.Set_to_Label(label_req, OriginIx2); + estate.Compute_All(false); + cout << spstate << endl; + cout << estate; + if (whichDSF == 'd') + cout << setprecision(16) << "estate.E - spstate.E = " << estate.E - spstate.E << "\tME = " << real(exp(ln_Density_ME(spstate, estate))) << endl; + else if (whichDSF == 'o') + cout << setprecision(16) << "estate.E - spstate.E = " << estate.E - spstate.E << "\tME = " << real(exp(ln_Psi_ME(estate, spstate))) << endl; + else if (whichDSF == 'g') + cout << setprecision(16) << "estate.E - spstate.E = " << estate.E - spstate.E << "\tME = " << real(exp(ln_Psi_ME(spstate, estate))) << endl; + + cout << "Another try ? (1 == yes, 0 == no)" << endl; + again = 1; + cin >> again; + } while (again != 0); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_Data_Daemon.cc b/src/EXECS/LiebLin_Data_Daemon.cc new file mode 100644 index 0000000..ea6e085 --- /dev/null +++ b/src/EXECS/LiebLin_Data_Daemon.cc @@ -0,0 +1,120 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Data_Daemon.cc + +Purpose: Produces sets of data files for correlations. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 11) { // provide some info + + cout << endl << "Welcome to ABACUS\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_Data_Daemon executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int_max \t\t Largest value of the interaction parameter: use positive real values only" << endl; + cout << "int Nc \t\t number of steps in interaction value" << endl; + cout << "int cfact \t\t dividing factor (each new interaction value if 1/cfact times the previous)" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int Max_Hrs \t\t Allowed computational time: (in hours)" << endl; + } + + else { // (argc == 10), correct nr of arguments + int ia = 1; + char whichDSF = *argv[ia++]; + DP c_int_max = atof(argv[ia++]); + int Nc = atoi(argv[ia++]); + int cfact = atoi(argv[ia++]); + DP L = atof(argv[ia++]); + int N = atoi(argv[ia++]); + int iKmin = atoi(argv[ia++]); + int iKmax = atoi(argv[ia++]); + DP kBT = atof(argv[ia++]); + int Max_Hrs = atoi(argv[ia++]); + + // Individual computations are split into chuncks of Max_Hrs/(Nc * 4) + //int Max_Secs = (Max_Hrs * 900)/Nc; + int Max_Secs = (Max_Hrs * 2700)/Nc; // to minimize wrapping up & restarting time + + cout << "Data daemon will use chunks of " << Max_Secs << " seconds." << endl; + + //clock_t StartTime = clock(); + double StartTime = omp_get_wtime(); + + //clock_t ActualTime = StartTime; + double ActualTime = omp_get_wtime(); + + DP c_int; + DP target_sumrule = 1.0; + + //while (double(ActualTime - StartTime)/CLOCKS_PER_SEC < double(3600 * Max_Hrs - Max_Secs)) { + while (ActualTime - StartTime < double(3600 * Max_Hrs - Max_Secs)) { + + Vect srsat(0.0, Nc); + Vect refine(false, Nc); + DP srmin = 1.0; + int icmin = 0; + + // Determine which correlation has the worst sum rule: + for (int ic = 0; ic < Nc; ++ic) { + c_int = c_int_max/pow(cfact, ic); + stringstream SRC_stringstream; string SRC_string; + Data_File_Name (SRC_stringstream, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + SRC_stringstream << ".src"; + SRC_string = SRC_stringstream.str(); const char* SRC_Cstr = SRC_string.c_str(); + + fstream srcfile; + srcfile.open(SRC_Cstr, fstream::in); + if (srcfile.fail()) { + srsat[ic] = 0.0; + refine[ic] = false; + } + else { + srcfile >> srsat[ic]; + refine[ic] = true; + } + if (srsat[ic] < srmin) { + srmin = srsat[ic]; + icmin = ic; + } + srcfile.close(); + + } // for ic + + cout << "srsat min found: " << srmin << "\t for c = " << c_int_max/pow(cfact, icmin) << ". Now refining this." + //<< " Time left: " << 3600* Max_Hrs - (ActualTime - StartTime)/CLOCKS_PER_SEC << " seconds." << endl; + << " Time left: " << 3600* Max_Hrs - (ActualTime - StartTime) << " seconds." << endl; + + // Improve the icmin calculation by one chunk: + Scan_LiebLin (whichDSF, c_int_max/pow(cfact, icmin), L, N, iKmin, iKmax, kBT, Max_Secs, target_sumrule, refine[icmin]); + + //ActualTime = clock(); + ActualTime = omp_get_wtime(); + + } // while there is time + + cout << "Wrapping up, time's up." << endl; + + } // else if arguments given OK + + return(0); +} + diff --git a/src/EXECS/LiebLin_Data_Daemon_Nscaling.cc b/src/EXECS/LiebLin_Data_Daemon_Nscaling.cc new file mode 100644 index 0000000..fae157d --- /dev/null +++ b/src/EXECS/LiebLin_Data_Daemon_Nscaling.cc @@ -0,0 +1,104 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Data_Daemon_Nscaling.cc + +Purpose: Produces sets of data files for correlations, increasing system size at fixed c and momentum window. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 9) { // provide some info + + cout << endl << "Welcome to ABACUS\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_Data_Daemon_Nscaling executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + //cout << "int Nc \t\t number of steps in interaction value" << endl; + cout << "int Nstep \t\t\t Steps to be taken in number of particles: use positive integer values only. Filling will be set to 1 (L == N)" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over (in units of N == Nstep): recommended values: 0 and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "int Max_minutes \t\t Allowed computational time: (in minutes)" << endl; + } + + else { // (argc == 10), correct nr of arguments + int ia = 1; + char whichDSF = *argv[ia++]; + DP c_int = atof(argv[ia++]); + int Nstep = atoi(argv[ia++]); + int iKmin_Nstep = atoi(argv[ia++]); + int iKmax_Nstep = atoi(argv[ia++]); + DP kBT = atof(argv[ia++]); + DP target_sumrule = atof(argv[ia++]); + int Max_minutes = atoi(argv[ia++]); + + + //clock_t StartTime = clock(); + double StartTime = omp_get_wtime(); + + //clock_t ActualTime = StartTime; + double ActualTime = omp_get_wtime(); + + int Secs_left = 60* Max_minutes; + + int iN = 0; + + while (Secs_left > 0) { + + iN += 1; + int N = Nstep * iN; + DP L = N; + int iKmin = iKmin_Nstep * iN; + int iKmax = iKmax_Nstep * iN; + DP srsat = 0.0; + bool refine = false; + + stringstream SRC_stringstream; string SRC_string; + Data_File_Name (SRC_stringstream, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + SRC_stringstream << ".src"; + SRC_string = SRC_stringstream.str(); const char* SRC_Cstr = SRC_string.c_str(); + + fstream srcfile; + srcfile.open(SRC_Cstr, fstream::in); + if (srcfile.fail()) { + srsat = 0.0; + refine = false; + } + else { + srcfile >> srsat; + refine = true; + } + srcfile.close(); + + // Improve the icmin calculation by one chunk: + Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, Secs_left, target_sumrule, refine); + + ActualTime = clock(); + + //Secs_left = int(60*Max_minutes - double(ActualTime - StartTime)/CLOCKS_PER_SEC); + Secs_left = int(60*Max_minutes - (ActualTime - StartTime)); + + cout << "Done with N = " << N << ". Time left = " << Secs_left << " seconds." << endl; + + } // while there is time + + } // else if arguments given OK + + return(0); +} + diff --git a/src/EXECS/LiebLin_Fourier_ssf_to_Qsqx.cc b/src/EXECS/LiebLin_Fourier_ssf_to_Qsqx.cc new file mode 100644 index 0000000..8d90882 --- /dev/null +++ b/src/EXECS/LiebLin_Fourier_ssf_to_Qsqx.cc @@ -0,0 +1,149 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Fourier_to_Qsqx.cc + +Purpose: Compute the Q(x)^2 expectation value for LiebLin, where Q(x) = \int_0^x dx \rho(x). + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 9) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_Fourier_to_Qsqx executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers scanned over" << endl; + cout << "DP kBT \t\t Temperature" << endl; + cout << "int Npts_x Number of points in space for the Fourier transform" << endl; + } + + else { // (argc == 9), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + int Npts_x = atoi(argv[8]); + // Force Npts_x + //Npts_x = L; + + if (whichDSF != 'd') JSCerror("Must use whichDSF == d in LiebLin_Fourier_ssf_to_Qsqx"); + + stringstream filenameprefix; + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + string prefix = filenameprefix.str(); + + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + ifstream RAW_infile; + RAW_infile.open(RAW_Cstr); + if (RAW_infile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("Could not open RAW_infile... "); + } + + // We also read the f-sumrule file, to correct for missing intensity. + stringstream FSR_stringstream; string FSR_string; + FSR_stringstream << prefix << ".fsr"; + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + ifstream FSR_infile; + FSR_infile.open(FSR_Cstr); + if (FSR_infile.fail()) { + cout << FSR_Cstr << endl; + JSCerror("Could not open FSR_infile... "); + } + + stringstream SFT_stringstream; string SFT_string; + SFT_stringstream << prefix << ".Qsqx"; + SFT_string = SFT_stringstream.str(); const char* SFT_Cstr = SFT_string.c_str(); + ofstream SFT_outfile; + SFT_outfile.open(SFT_Cstr); + if (SFT_outfile.fail()) JSCerror("Could not open SFT_outfile... "); + + // First compute the static structure factor from the RAW data: + + Vect_DP SSF(0.0, iKmax - iKmin + 1); + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + while (RAW_infile.peek() != EOF) { + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (iK >= iKmin && iK <= iKmax) { + SSF[iK - iKmin] += FF * FF; + } + } + RAW_infile.close(); + + // Reset proper normalization: + DP normalization = twoPI * L; + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) SSF[iK] *= normalization/twoPI; // twoPI from integral over omega + + + // Now define real-space coordinates: between 0 and L/2 + Vect_DP xlattice(Npts_x); + for (int i = 0; i < Npts_x; ++i) xlattice[i] = (i + 0.5) * 0.5* L/Npts_x; + + // Now the correlation at x: + Vect_DP FT(0.0, Npts_x); + + DP pioverL = PI/L; + + // Fourier transform: + for (int ix = 0; ix < Npts_x; ++ix) { + for (int iK = 1; iK <= iKmax; ++iK) { + FT[ix] += SSF[iK - iKmin] * pow(sin(pioverL * iK * xlattice[ix])/iK, 2.0); + } + // Reset proper normalization: 1/L from space FT, + FT[ix] *= 2.0*L/(PI * PI); + + // Outside of window iKmin, iKmax, we take the DSF to be a constant with delta function + // at free energy k^2, so DSF = 2\pi N/L \delta(\omega - k^2) (to fit f-sumrule) + // so SSF becomes N/L. + // We thus need to correct above by adding + // \frac{1}{L} \sum_{-\infty}^{iKmin - 1} SSF e^{ikx} + \frac{1}{L} \sum_{iKmax + 1}^\infty SSF e^{ikx} + // Resumming carefully: + //FTre[ix] += (sin(twopioverL * (iKmin - 0.5) * xlattice[ix]) - sin(twopioverL * (iKmax + 0.5) * xlattice[ix])) + //* N/(2.0 * L*L * sin(PI * xlattice[ix]/L)); + //FTim[ix] += (-cos(twopioverL * (iKmin - 0.5) * xlattice[ix]) + cos(twopioverL * (iKmax + 0.5) * xlattice[ix])) + //* N/(2.0 * L*L * sin(PI * xlattice[ix]/L)); + } + + + // Output to file: + for (int ix = 0; ix < Npts_x; ++ix) { + if (ix > 0) SFT_outfile << endl; + //SFT_outfile << xlattice[ix] << "\t" << FTre[ix] << "\t" << FTim[ix] << "\t" << FTreavg[ix] << "\t" << FTimavg[ix]; + SFT_outfile << xlattice[ix]/L << "\t" << FT[ix]; + } + + SFT_outfile.close(); + } + + return(0); + +} diff --git a/src/EXECS/LiebLin_Fourier_to_t_equal_x.cc b/src/EXECS/LiebLin_Fourier_to_t_equal_x.cc new file mode 100644 index 0000000..a689bb1 --- /dev/null +++ b/src/EXECS/LiebLin_Fourier_to_t_equal_x.cc @@ -0,0 +1,132 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Fourier_to_t_equal_x.cc + +Purpose: Fourier transform to static space correlator for LiebLin. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_Fourier_to_x_equal_t executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers scanned over" << endl; + cout << "DP kBT \t\t Temperature" << endl; + cout << "int Npts_t \t Number of points in time for the Fourier transform" << endl; + cout << "DP t_max \t Max time to be used" << endl; + } + + else { // (argc == 10), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + int Npts_t = atoi(argv[8]); + DP t_max = atof(argv[9]); + + // Momentum business: use symmetry + if (iKmin != 0) JSCerror("LiebLin_Fourier_to_t_equal_x only implemented for raw files with iKmin == 0."); + + stringstream filenameprefix; + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + string prefix = filenameprefix.str(); + + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + ifstream RAW_infile; + RAW_infile.open(RAW_Cstr); + if (RAW_infile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("Could not open RAW_infile... "); + } + + // We also read the f-sumrule file, to correct for missing intensity. + stringstream FSR_stringstream; string FSR_string; + FSR_stringstream << prefix << ".fsr"; + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + ifstream FSR_infile; + FSR_infile.open(FSR_Cstr); + if (FSR_infile.fail()) { + cout << FSR_Cstr << endl; + JSCerror("Could not open FSR_infile... "); + } + + stringstream SFT_stringstream; string SFT_string; + SFT_stringstream << prefix << ".tft"; + SFT_string = SFT_stringstream.str(); const char* SFT_Cstr = SFT_string.c_str(); + ofstream SFT_outfile; + SFT_outfile.open(SFT_Cstr); + if (SFT_outfile.fail()) JSCerror("Could not open TFT_outfile... "); + + // First compute the static structure factor from the RAW data: + + Vect_DP TSF(0.0, Npts_t); + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + // Now define time coordinates: between 0 and t_max + Vect_DP tlattice(Npts_t); + for (int i = 0; i < Npts_t; ++i) tlattice[i] = (i + 0.5) * t_max/Npts_t; + + // Now the correlation at t: + Vect > FT(0.0, Npts_t); + Vect_DP FTre(0.0, Npts_t); + Vect_DP FTim(0.0, Npts_t); + + while (RAW_infile.peek() != EOF) { + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (iK == 0) + for (int it = 0; it < Npts_t; ++it) + FT[it] += FF * FF * exp(II * omega * tlattice[it]); + else + for (int it = 0; it < Npts_t; ++it) + FT[it] += 2.0 * FF * FF * exp(II * omega * tlattice[it]); + } + RAW_infile.close(); + + // Reset proper normalization: + for (int it = 0; it < Npts_t; ++it) { + FTre[it] = real(FT[it]); + FTim[it] = imag(FT[it]); + } + + // Output to file: + for (int it = 0; it < Npts_t; ++it) { + if (it > 0) SFT_outfile << endl; + SFT_outfile << tlattice[it] << "\t" << FTre[it] << "\t" << FTim[it]; + } + + SFT_outfile.close(); + } + + return(0); + +} diff --git a/src/EXECS/LiebLin_Fourier_to_t_equal_x_from_RAW.cc b/src/EXECS/LiebLin_Fourier_to_t_equal_x_from_RAW.cc new file mode 100644 index 0000000..8a11b3b --- /dev/null +++ b/src/EXECS/LiebLin_Fourier_to_t_equal_x_from_RAW.cc @@ -0,0 +1,118 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Fourier_to_t_equal_x.cc + +Purpose: Fourier transform to static space correlator for LiebLin. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_Fourier_to_x_equal_t executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers scanned over" << endl; + //cout << "DP kBT \t\t Temperature" << endl; + cout << "RAW file name" << endl; + cout << "int Npts_t \t Number of points in time for the Fourier transform" << endl; + cout << "DP t_max \t Max time to be used" << endl; + } + + else { // (argc == 10), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + //DP kBT = atof(argv[7]); + char* rawfilename = argv[7]; + int Npts_t = atoi(argv[8]); + DP t_max = atof(argv[9]); + + // Momentum business: use symmetry + if (iKmin != 0) JSCerror("LiebLin_Fourier_to_t_equal_x only implemented for raw files with iKmin == 0."); + + ifstream RAW_infile; + //RAW_infile.open(RAW_Cstr); + RAW_infile.open(rawfilename); + if (RAW_infile.fail()) { + //cout << RAW_Cstr << endl; + cout << rawfilename << endl; + JSCerror("Could not open RAW_infile... "); + } + + stringstream SFT_stringstream; string SFT_string; + SFT_stringstream << rawfilename << "_tft"; + SFT_string = SFT_stringstream.str(); const char* SFT_Cstr = SFT_string.c_str(); + ofstream SFT_outfile; + SFT_outfile.open(SFT_Cstr); + if (SFT_outfile.fail()) JSCerror("Could not open TFT_outfile... "); + + // First compute the static structure factor from the RAW data: + + Vect_DP TSF(0.0, Npts_t); + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + // Now define time coordinates: between 0 and t_max + Vect_DP tlattice(Npts_t); + for (int i = 0; i < Npts_t; ++i) tlattice[i] = (i + 0.5) * t_max/Npts_t; + + // Now the correlation at t: + Vect > FT(0.0, Npts_t); + Vect_DP FTre(0.0, Npts_t); + Vect_DP FTim(0.0, Npts_t); + + while (RAW_infile.peek() != EOF) { + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (iK == 0) + for (int it = 0; it < Npts_t; ++it) + FT[it] += FF * FF * exp(II * omega * tlattice[it]); + else + for (int it = 0; it < Npts_t; ++it) + FT[it] += 2.0 * FF * FF * exp(II * omega * tlattice[it]); + } + RAW_infile.close(); + + // Reset proper normalization: + for (int it = 0; it < Npts_t; ++it) { + FTre[it] = real(FT[it]); + FTim[it] = imag(FT[it]); + } + + // Output to file: + for (int it = 0; it < Npts_t; ++it) { + if (it > 0) SFT_outfile << endl; + SFT_outfile << tlattice[it] << "\t" << FTre[it] << "\t" << FTim[it]; + } + + SFT_outfile.close(); + } + + return(0); + +} diff --git a/src/EXECS/LiebLin_Fourier_to_x_equal_t.cc b/src/EXECS/LiebLin_Fourier_to_x_equal_t.cc new file mode 100644 index 0000000..6c473c1 --- /dev/null +++ b/src/EXECS/LiebLin_Fourier_to_x_equal_t.cc @@ -0,0 +1,191 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Fourier_to_x_equal_t.cc + +Purpose: Fourier transform to static space correlator for LiebLin. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 9) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_Fourier_to_x_equal_t executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers scanned over" << endl; + cout << "DP kBT \t\t Temperature" << endl; + cout << "int Npts_x Number of points in space for the Fourier transform" << endl; + } + + else { // (argc == 9), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + int Npts_x = atoi(argv[8]); + // Force Npts_x + //Npts_x = L; + + stringstream filenameprefix; + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + string prefix = filenameprefix.str(); + + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + ifstream RAW_infile; + RAW_infile.open(RAW_Cstr); + if (RAW_infile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("Could not open RAW_infile... "); + } + + // We also read the f-sumrule file, to correct for missing intensity. + stringstream FSR_stringstream; string FSR_string; + FSR_stringstream << prefix << ".fsr"; + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + ifstream FSR_infile; + FSR_infile.open(FSR_Cstr); + if (FSR_infile.fail()) { + cout << FSR_Cstr << endl; + JSCerror("Could not open FSR_infile... "); + } + + stringstream SFT_stringstream; string SFT_string; + SFT_stringstream << prefix << ".sft"; + SFT_string = SFT_stringstream.str(); const char* SFT_Cstr = SFT_string.c_str(); + ofstream SFT_outfile; + SFT_outfile.open(SFT_Cstr); + if (SFT_outfile.fail()) JSCerror("Could not open SFT_outfile... "); + + // First compute the static structure factor from the RAW data: + + Vect_DP SSF(0.0, iKmax - iKmin + 1); + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + while (RAW_infile.peek() != EOF) { + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (iK >= iKmin && iK <= iKmax) { + SSF[iK - iKmin] += FF * FF; + } + } + RAW_infile.close(); + + // Reset proper normalization: + DP normalization = twoPI * L; + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) SSF[iK] *= normalization/twoPI; // twoPI from integral over omega + + // We now refine the SSF in the following way. + // First, we read off the f-sumrule saturation from the FSR file. + // Then, we put back the missing weight, assuming that it lives + // on the free k^2 dispersion relation (so the DSF is then simply 2\pi N/L \delta(\omega - k^2)). + + Vect_DP FSRsaturated(0.0, iKmax - iKmin + 1); + int dummyint; + DP dummy; + for (int i = iKmin; i <= iKmax; ++i) + FSR_infile >> dummyint >> FSRsaturated[i - iKmin] >> dummy; + + FSR_infile.close(); + + // Now correct the SSF by the missing piece: + //for (int iK = iKmin; iK <= iKmax; ++iK) + //SSF[iK - iKmin] += (1.0 - FSRsaturated[iK - iKmin]) * N/L; + + //cout << FSRsaturated << endl; + + //cout << SSF << endl; + + // Now define real-space coordinates: between 0 and L + Vect_DP xlattice(Npts_x); + for (int i = 0; i < Npts_x; ++i) xlattice[i] = (i + 0.5) * L/Npts_x; + + // Now the correlation at x: + Vect_DP FTre(0.0, Npts_x); + Vect_DP FTim(0.0, Npts_x); + + DP twopioverL = twoPI/L; + + // Fourier transform: + for (int ix = 0; ix < Npts_x; ++ix) { + for (int iK = iKmin; iK <= iKmax; ++iK) { + FTre[ix] += SSF[iK - iKmin] * cos(twopioverL * iK * xlattice[ix]); + FTim[ix] += SSF[iK - iKmin] * sin(twopioverL * iK * xlattice[ix]); + } + // Reset proper normalization: 1/L from space FT, + FTre[ix] /= L; + FTim[ix] /= L; + + // Outside of window iKmin, iKmax, we take the DSF to be a constant with delta function + // at free energy k^2, so DSF = 2\pi N/L \delta(\omega - k^2) (to fit f-sumrule) + // so SSF becomes N/L. + // We thus need to correct above by adding + // \frac{1}{L} \sum_{-\infty}^{iKmin - 1} SSF e^{ikx} + \frac{1}{L} \sum_{iKmax + 1}^\infty SSF e^{ikx} + // Resumming carefully: + if (whichDSF == 'd') { + FTre[ix] += (sin(twopioverL * (iKmin - 0.5) * xlattice[ix]) - sin(twopioverL * (iKmax + 0.5) * xlattice[ix])) + * N/(2.0 * L*L * sin(PI * xlattice[ix]/L)); + FTim[ix] += (-cos(twopioverL * (iKmin - 0.5) * xlattice[ix]) + cos(twopioverL * (iKmax + 0.5) * xlattice[ix])) + * N/(2.0 * L*L * sin(PI * xlattice[ix]/L)); + } + } + + // Since iKmax and iKmin are finite, we need to average over an interval of + // deltax such that (2\pi/L) iKmax deltax = 2\pi, with deltax == deltaix * L/Npts_x + // so deltaix = (Npts_x/L) * (L/iKmax) + /* + int deltaix = 0*int(Npts_x/(2.0*iKmax)); + cout << "deltaix = " << deltaix << endl; + + Vect_DP FTreavg(0.0, Npts_x); + Vect_DP FTimavg(0.0, Npts_x); + for (int ix = 0; ix < Npts_x; ++ix) { + for (int ix2 = -deltaix; ix2 < deltaix; ++ix2) { + FTreavg[ix] += FTre[JSC::min(JSC::max(0, ix + ix2), Npts_x - 1)]; + FTimavg[ix] += FTim[JSC::min(JSC::max(0, ix + ix2), Npts_x - 1)]; + } + FTreavg[ix] /= (2*deltaix + 1); + FTimavg[ix] /= (2*deltaix + 1); + } + */ + if (whichDSF == 'd') cout << "g2(0) = dE0_dc/L = " << LiebLin_dE0_dc (c_int, L, N)/L << "\t" << LiebLin_dE0_dc (c_int, 2.0*L, 2*N)/(2.0*L) << endl; + + // Output to file: + for (int ix = 0; ix < Npts_x; ++ix) { + if (ix > 0) SFT_outfile << endl; + //SFT_outfile << xlattice[ix] << "\t" << FTre[ix] << "\t" << FTim[ix] << "\t" << FTreavg[ix] << "\t" << FTimavg[ix]; + SFT_outfile << xlattice[ix] << "\t" << FTre[ix] << "\t" << FTim[ix]; + } + + SFT_outfile.close(); + } + + return(0); + +} diff --git a/src/EXECS/LiebLin_Fourier_to_x_equal_t_from_RAW.cc b/src/EXECS/LiebLin_Fourier_to_x_equal_t_from_RAW.cc new file mode 100644 index 0000000..2c49b25 --- /dev/null +++ b/src/EXECS/LiebLin_Fourier_to_x_equal_t_from_RAW.cc @@ -0,0 +1,147 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Fourier_to_x_equal_t.cc + +Purpose: Fourier transform to static space correlator for LiebLin. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 9) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_Fourier_to_x_equal_t executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers scanned over" << endl; + //cout << "DP kBT \t\t Temperature" << endl; + cout << "RAW file name" << endl; + cout << "int Npts_x Number of points in space for the Fourier transform" << endl; + } + + else { // (argc == 9), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + //DP kBT = atof(argv[7]); + char* rawfilename = argv[7]; + int Npts_x = atoi(argv[8]); + // Force Npts_x + //Npts_x = L; + + //stringstream filenameprefix; + //Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + //string prefix = filenameprefix.str(); + + //stringstream RAW_stringstream; string RAW_string; + //RAW_stringstream << prefix << ".raw"; + //RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + ifstream RAW_infile; + //RAW_infile.open(RAW_Cstr); + RAW_infile.open(rawfilename); + if (RAW_infile.fail()) { + //cout << RAW_Cstr << endl; + cout << rawfilename << endl; + JSCerror("Could not open RAW_infile... "); + } + + // Define the output file name: use the RAW file name but with different suffix + + stringstream SFT_stringstream; string SFT_string; + SFT_stringstream << rawfilename << "_sft"; + SFT_string = SFT_stringstream.str(); const char* SFT_Cstr = SFT_string.c_str(); + ofstream SFT_outfile; + SFT_outfile.open(SFT_Cstr); + if (SFT_outfile.fail()) JSCerror("Could not open SFT_outfile... "); + + // First compute the static structure factor from the RAW data: + + Vect_DP SSF(0.0, iKmax - iKmin + 1); + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + while (RAW_infile.peek() != EOF) { + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (iK >= iKmin && iK <= iKmax) { + SSF[iK - iKmin] += FF * FF; + } + } + RAW_infile.close(); + + // Reset proper normalization: + DP normalization = twoPI * L; + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) SSF[iK] *= normalization/twoPI; // twoPI from integral over omega + + + // Now define real-space coordinates: between 0 and L + Vect_DP xlattice(Npts_x); + for (int i = 0; i < Npts_x; ++i) xlattice[i] = (i + 0.5) * L/Npts_x; + + // Now the correlation at x: + Vect_DP FTre(0.0, Npts_x); + Vect_DP FTim(0.0, Npts_x); + + DP twopioverL = twoPI/L; + + // Fourier transform: + for (int ix = 0; ix < Npts_x; ++ix) { + for (int iK = iKmin; iK <= iKmax; ++iK) { + FTre[ix] += SSF[iK - iKmin] * cos(twopioverL * iK * xlattice[ix]); + FTim[ix] += SSF[iK - iKmin] * sin(twopioverL * iK * xlattice[ix]); + } + // Reset proper normalization: 1/L from space FT, + FTre[ix] /= L; + FTim[ix] /= L; + + // Outside of window iKmin, iKmax, we take the DSF to be a constant with delta function + // at free energy k^2, so DSF = 2\pi N/L \delta(\omega - k^2) (to fit f-sumrule) + // so SSF becomes N/L. + // We thus need to correct above by adding + // \frac{1}{L} \sum_{-\infty}^{iKmin - 1} SSF e^{ikx} + \frac{1}{L} \sum_{iKmax + 1}^\infty SSF e^{ikx} + // Resumming carefully: + //if (whichDSF == 'd') { + //FTre[ix] += (sin(twopioverL * (iKmin - 0.5) * xlattice[ix]) - sin(twopioverL * (iKmax + 0.5) * xlattice[ix])) + // * N/(2.0 * L*L * sin(PI * xlattice[ix]/L)); + //FTim[ix] += (-cos(twopioverL * (iKmin - 0.5) * xlattice[ix]) + cos(twopioverL * (iKmax + 0.5) * xlattice[ix])) + // * N/(2.0 * L*L * sin(PI * xlattice[ix]/L)); + //} + } + + + // Output to file: + for (int ix = 0; ix < Npts_x; ++ix) { + if (ix > 0) SFT_outfile << endl; + //SFT_outfile << xlattice[ix] << "\t" << FTre[ix] << "\t" << FTim[ix] << "\t" << FTreavg[ix] << "\t" << FTimavg[ix]; + SFT_outfile << xlattice[ix] << "\t" << FTre[ix] << "\t" << FTim[ix]; + } + + SFT_outfile.close(); + } + + return(0); + +} diff --git a/src/EXECS/LiebLin_Moses_tester.cc b/src/EXECS/LiebLin_Moses_tester.cc new file mode 100644 index 0000000..4ca4c04 --- /dev/null +++ b/src/EXECS/LiebLin_Moses_tester.cc @@ -0,0 +1,79 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_DSF_tester.cc + +Purpose: allows for Ix2 manipulations (user-prompted) for LiebLin gas + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 8) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_DSF_tester executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int Nl \t\t\t Number of particles in left Fermi sea (Nr is then N - Nl)" << endl; + cout << "int DIl \t\t shift of left sea as compared to its ground state position" << endl; + cout << "int DIr \t\t shift of right sea as compared to its ground state position" << endl; + } + + else { // (argc == 6), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int Nl = atoi(argv[5]); + //int Nr = N - Nl; + int DIl = atoi(argv[6]); + int DIr = atoi(argv[7]); + + if (whichDSF != 'd') JSCerror("Other options not implemented yet in LiebLin_Moses_tester"); + + // Define the Moses state: + LiebLin_Bethe_State MosesState (c_int, L, N); + + // Split the sea: + for (int i = 0; i < Nl; ++i) MosesState.Ix2[i] -= 2 * DIl; + for (int i = Nl; i < N; ++i) MosesState.Ix2[i] += 2 * DIr; + + MosesState.Compute_All (true); + + LiebLin_Bethe_State estate = MosesState; + cout << MosesState; + + int Ix2old, Ix2new; + int again = 0; + do { + cout << "Substitute Ix2: which for which ?" << endl; + cin >> Ix2old >> Ix2new; + for (int i = 0; i < estate.N; ++i) if (estate.Ix2[i] == Ix2old) estate.Ix2[i] = Ix2new; + estate.Compute_All(false); + cout << estate; + cout << setprecision(16) << "omega = " << estate.E - MosesState.E << "\t" << exp(real(ln_Density_ME(MosesState, estate))) << endl; + cout << "Another try ? (0 == no)" << endl; + cin >> again; + } while (again != 0); + + } + + return(0); +} + diff --git a/src/EXECS/LiebLin_RAW_File_Stats.cc b/src/EXECS/LiebLin_RAW_File_Stats.cc new file mode 100644 index 0000000..039cd4a --- /dev/null +++ b/src/EXECS/LiebLin_RAW_File_Stats.cc @@ -0,0 +1,134 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_RAW_File_stats.cc + +Purpose: Analyzes the distribution of matrix element values in a RAW file, + to help with optimization of the scanning procedure. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 9) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of LiebLin_RAW_File_Stats executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter: use positive real values only" << endl; + cout << "DP L \t\t\t Length of the system: use positive real values only" << endl; + cout << "int N \t\t\t Number of particles: use positive integer values only" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: -2*N and 2*N" << endl; + cout << "DP kBT \t\t Temperature (positive only of course)" << endl; + cout << "int Aggregate size" << endl; + } + + else { // (argc == 9), correct nr of arguments + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + int AgSize = atoi(argv[8]); + + if (AgSize < 2) JSCerror("Give an aggregate size > 1 in LiebLin_RAW_File_Stats."); + + stringstream RAW_stringstream; string RAW_string; + Data_File_Name (RAW_stringstream, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + RAW_stringstream << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + ifstream RAW_infile; + RAW_infile.open(RAW_Cstr); + if (RAW_infile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("Could not open RAW_infile... "); + } + + stringstream STAT_stringstream; string STAT_string; + Data_File_Name (STAT_stringstream, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + STAT_stringstream << ".stat"; + STAT_string = STAT_stringstream.str(); const char* STAT_Cstr = STAT_string.c_str(); + + ofstream STATfile; + STATfile.open(STAT_Cstr); + if (STATfile.fail()) { + cout << STAT_Cstr << endl; JSCerror("Could not open STATfile."); + } + + LiebLin_Bethe_State AveragingState = Canonical_Saddle_Point_State (c_int, L, N, whichDSF == 'Z' ? 0.0 : kBT); + + DP Chem_Pot = Chemical_Potential (AveragingState); + //DP sumrule_factor = Sumrule_Factor (whichDSF, AveragingState, Chem_Pot, fixed_iK, iKneeded); + Vect sumrule_factor(iKmax - iKmin + 1); + for (int ik = 0; ik < iKmax - iKmin + 1; ++ik) + sumrule_factor[ik] = Sumrule_Factor (whichDSF, AveragingState, Chem_Pot, ik, ik); + // Normalize by total number of considered momenta + DP correction_factor = 1.0/(iKmax - iKmin + 1); + if (iKmin <= 0 && iKmax >= 0 && whichDSF == 'd') { // correct for fact that RhoRho is 0 at k = 0 + sumrule_factor[0] = 0.0; + correction_factor = 1.0/(iKmax - iKmin); + } + + for (int ik = 0; ik < iKmax - iKmin + 1; ++ik) sumrule_factor[ik] *= correction_factor; + + DP omega; + int iK; + DP ME; + DP dev; + string label; + + int nread = 0; + + DP srcont = 0.0; + DP abssrcont = 0.0; + DP maxsrcont = 0.0; + DP totsrcont = 0.0; + DP accumulatedsrcont = 0.0; + int naccounted = 0; + + while (RAW_infile.peek() != EOF) { + + RAW_infile >> omega >> iK >> ME >> dev >> label; + nread++; + + if (iK >= iKmin && iK <= iKmax) { + srcont = omega * ME * ME * sumrule_factor[iK - iKmin]; + abssrcont = fabs(srcont); + maxsrcont = JSC::max(maxsrcont, abssrcont); + totsrcont += srcont; + accumulatedsrcont += srcont; + naccounted++; + } + + if (naccounted >= AgSize) { + STATfile << nread << "\t" << maxsrcont << "\t" << totsrcont/AgSize << "\t" << totsrcont/(AgSize * (maxsrcont > 0.0 ? maxsrcont : 1.0)) << "\t" << accumulatedsrcont << endl; + naccounted = 0; + maxsrcont = 0.0; + totsrcont = 0.0; + } + } + + RAW_infile.close(); + STATfile.close(); + } + + + return(0); +} + diff --git a/src/EXECS/LiebLin_TBA.cc b/src/EXECS/LiebLin_TBA.cc new file mode 100755 index 0000000..08ebe9f --- /dev/null +++ b/src/EXECS/LiebLin_TBA.cc @@ -0,0 +1,45 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_TBA.cc + +Purpose: solves the TBA equations for Lieb-Liniger + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +int main(int argc, const char* argv[]) +{ + + //if (argc != 7) JSCerror("Wrong number of arguments to 2CBG_ThLim executable. Use c(best to set to 1), nbar, ebar, req_diff, Max_Secs, bool Save_data (0 == false)."); + if (argc != 6) JSCerror("Wrong number of arguments. Use c(best to set to 1), mu, kBT, req_diff, Max_Secs"); + + DP c_int = atof(argv[1]); + DP mu = atof(argv[2]); + DP kBT = atof(argv[3]); + DP req_diff = atof(argv[4]); + int Max_Secs = atoi(argv[5]); + + if (c_int <= 0.0) JSCerror("Give a strictly positive c."); + if (kBT <= 0.0) JSCerror("Negative T ? Not for the LiebLin gas."); + if (Max_Secs < 10) JSCerror("Give more time."); + + //cout << "Read c_int = " << c_int << "\tmu = " << mu << "\tOmega = " << Omega << "\tkBT = " << kBT << "\tMax_Secs = " << Max_Secs << endl; + + LiebLin_TBA_Solution solution(c_int, mu, kBT, req_diff, Max_Secs); + + + cout << solution.nbar << "\t" << solution.ebar << "\t" << solution.sbar << "\t"; + + return(0); +} diff --git a/src/EXECS/LiebLin_TBA_fixed_nbar.cc b/src/EXECS/LiebLin_TBA_fixed_nbar.cc new file mode 100755 index 0000000..a69b3b8 --- /dev/null +++ b/src/EXECS/LiebLin_TBA_fixed_nbar.cc @@ -0,0 +1,44 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_TBA_fixed_nbar.cc + +Purpose: solves the TBA equations for Lieb-Liniger + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +int main(int argc, const char* argv[]) +{ + + //if (argc != 7) JSCerror("Wrong number of arguments to 2CBG_ThLim executable. Use c(best to set to 1), nbar, ebar, req_diff, Max_Secs, bool Save_data (0 == false)."); + if (argc != 6) JSCerror("Wrong number of arguments. Use c(best to set to 1), nbar, kBT, req_diff, Max_Secs"); + + DP c_int = atof(argv[1]); + DP nbar = atof(argv[2]); + DP kBT = atof(argv[3]); + DP req_diff = atof(argv[4]); + int Max_Secs = atoi(argv[5]); + + if (c_int <= 0.0) JSCerror("Give a strictly positive c."); + if (kBT <= 0.0) JSCerror("Negative T ? Not for the LiebLin gas."); + if (Max_Secs < 10) JSCerror("Give more time."); + + //cout << "Read c_int = " << c_int << "\tmu = " << mu << "\tOmega = " << Omega << "\tkBT = " << kBT << "\tMax_Secs = " << Max_Secs << endl; + + LiebLin_TBA_Solution solution = LiebLin_TBA_Solution_fixed_nbar (c_int, nbar, kBT, req_diff, Max_Secs); + + cout << solution.nbar << "\t" << solution.ebar << "\t" << solution.sbar << "\t"; + + return(0); +} diff --git a/src/EXECS/LiebLin_TBA_fixed_nbar_ebar.cc b/src/EXECS/LiebLin_TBA_fixed_nbar_ebar.cc new file mode 100755 index 0000000..e40d15e --- /dev/null +++ b/src/EXECS/LiebLin_TBA_fixed_nbar_ebar.cc @@ -0,0 +1,43 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_TBA_fixed_nbar_ebar.cc + +Purpose: solves the TBA equations for Lieb-Liniger + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +int main(int argc, const char* argv[]) +{ + + if (argc != 7) JSCerror("Wrong number of arguments. Use c(best to set to 1), nbar, ebar, req_diff, Max_Secs, bool Save_data (0 == false)."); + + DP c_int = atof(argv[1]); + DP nbar = atof(argv[2]); + DP ebar = atof(argv[3]); + DP req_diff = atof(argv[4]); + int Max_Secs = atoi(argv[5]); + bool Save_data = bool(atoi(argv[6])); + + if (c_int <= 0.0) JSCerror("Give a strictly positive c."); + if (Max_Secs < 10) JSCerror("Give more time."); + + //cout << "Read c_int = " << c_int << "\tmu = " << mu << "\tOmega = " << Omega << "\tkBT = " << kBT << "\tMax_Secs = " << Max_Secs << endl; + + LiebLin_TBA_Solution solution = LiebLin_TBA_Solution_fixed_nbar_ebar(c_int, nbar, ebar, req_diff, Max_Secs); + + cout << solution.nbar << "\t" << solution.ebar << "\t" << solution.sbar << "\t"; + + return(0); +} diff --git a/src/EXECS/ODSLF_DSF.cc b/src/EXECS/ODSLF_DSF.cc new file mode 100644 index 0000000..fe27258 --- /dev/null +++ b/src/EXECS/ODSLF_DSF.cc @@ -0,0 +1,60 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: ODSLF_DSF.cc + +Purpose: main function for ABACUS++ for spinless fermions related to Heisenberg spin-1/2 chain + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 10) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << "Usage of ODSLF_DSF executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: m for S- S+, z for Sz Sz, p for S+ S-." << endl; + cout << "DP Delta \t\t Value of the anisotropy: use positive real values only" << endl; + cout << "int N \t\t\t Length (number of sites) of the system: use positive even integer values only" << endl; + cout << "int M \t\t\t Number of down spins: use positive integer values between 1 and N/2" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers to scan over: recommended values: 0 and N" << endl; + cout << "int Max_Secs \t\t Allowed computational time: (in seconds)" << endl; + cout << "DP target_sumrule \t sumrule saturation you're satisfied with" << endl; + cout << "bool refine \t\t Is this a refinement of a earlier calculations ? (0 == false, 1 == true)" << endl; + cout << endl << "EXAMPLE: " << endl << endl; + cout << "ODSLF_DSF z 1.0 100 40 0 100 600 1.0 0" << endl << endl; + } + + else if (argc == 10) { // !fixed_iK + char whichDSF = *argv[1]; + DP Delta = atof(argv[2]); + int N = atoi(argv[3]); + int M = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + int Max_Secs = atoi(argv[7]); + DP target_sumrule = atof(argv[8]); + bool refine = (atoi(argv[9]) == 1); + + Scan_ODSLF (whichDSF, Delta, N, M, iKmin, iKmax, Max_Secs, target_sumrule, refine, 0, 1); + } + + + else JSCerror("Wrong number of arguments to ODSLF_DSF executable."); + + return(0); +} + diff --git a/src/EXECS/Produce_Sorted_RAW_File.cc b/src/EXECS/Produce_Sorted_RAW_File.cc new file mode 100644 index 0000000..3175a41 --- /dev/null +++ b/src/EXECS/Produce_Sorted_RAW_File.cc @@ -0,0 +1,37 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Sort_RAW_File.cc + +Purpose: produce a sorted .raw file. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 4) { + cout << "Arguments needed: rawfile, whichDSF, whichsorting." << endl; + JSCerror(""); + } + + const char* rawfilename = argv[1]; + char whichDSF = *argv[2]; + char whichsorting = *argv[3]; + + + //cout << rawfilename << "\t" << whichDSF << "\t" << whichsorting << endl; + Sort_RAW_File (rawfilename, whichsorting, whichDSF); + + return(0); +} diff --git a/src/EXECS/RAW_File_Stats.cc b/src/EXECS/RAW_File_Stats.cc new file mode 100644 index 0000000..b0ed325 --- /dev/null +++ b/src/EXECS/RAW_File_Stats.cc @@ -0,0 +1,99 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: RAW_File_stats.cc + +Purpose: Analyzes the distribution of matrix element values in a RAW file, + to help with optimization of the scanning procedure. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + + if (argc != 3) { // provide some info + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of RAW_File_Stats executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << endl << "string RAW file name" << endl; + cout << "int Aggregate size" << endl; + } + + else { // correct nr of arguments + const char* rawfilename = argv[1]; + int AgSize = atoi(argv[2]); + + if (AgSize < 2) JSCerror("Give an aggregate size > 1 in LiebLin_RAW_File_Stats."); + + ifstream RAW_infile; + RAW_infile.open(rawfilename); + if (RAW_infile.fail()) { + cout << rawfilename << endl; + JSCerror("Could not open RAW_infile... "); + } + + stringstream STAT_stringstream; string STAT_string; + STAT_stringstream << rawfilename << "_AgS_" << AgSize << "_stat"; + STAT_string = STAT_stringstream.str(); const char* STAT_Cstr = STAT_string.c_str(); + + ofstream STATfile; + STATfile.open(STAT_Cstr); + if (STATfile.fail()) { + cout << STAT_Cstr << endl; JSCerror("Could not open STATfile."); + } + + DP omega; + int iK; + DP ME; + DP dev; + string label; + + int nread = 0; + + DP srcont = 0.0; + DP abssrcont = 0.0; + DP maxsrcont = 0.0; + DP totsrcont = 0.0; + DP accumulatedsrcont = 0.0; + int naccounted = 0; + + while (RAW_infile.peek() != EOF) { + + RAW_infile >> omega >> iK >> ME >> dev >> label; + nread++; + + srcont = ME * ME; + abssrcont = fabs(srcont); + maxsrcont = JSC::max(maxsrcont, abssrcont); + totsrcont += srcont; + accumulatedsrcont += srcont; + naccounted++; + + if (naccounted >= AgSize) { + STATfile << nread << "\t" << maxsrcont << "\t" << totsrcont/AgSize << "\t" << totsrcont/(AgSize * (maxsrcont > 0.0 ? maxsrcont : 1.0)) << "\t" << accumulatedsrcont << endl; + naccounted = 0; + maxsrcont = 0.0; + totsrcont = 0.0; + } + } + + RAW_infile.close(); + STATfile.close(); + } + + + return(0); +} + diff --git a/src/EXECS/Smoothen_Heis_DSF.cc b/src/EXECS/Smoothen_Heis_DSF.cc new file mode 100644 index 0000000..98b6422 --- /dev/null +++ b/src/EXECS/Smoothen_Heis_DSF.cc @@ -0,0 +1,90 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Smoothen_Heis_DSF.cc + +Purpose: produces .dsf and .ssf files from a .raw file + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 13 && argc != 14) { // Print out instructions + //if (strcmp(argv[1],"help") == 0) { // Output some instructions + cout << "Usage of Smoothen_Heis_DSF executable: " << endl << endl; + cout << "Provide arguments using one of the following options:" << endl << endl; + cout << "1) (for general momenta) whichDSF Delta N M iKmin iKmax DiK kBT ommin ommax Nom gwidth" << endl << endl; + cout << "2) (for fixed momentum) whichDSF Delta N M iKneeded ommin ommax Nom gwidth" << endl << endl; + //else JSCerror("Incomprehensible arguments in Smoothen_Heis_DSF executable."); + } + + else if (argc == 13) { // !fixed_iK + char whichDSF = *argv[1]; + DP Delta = atof(argv[2]); + int N = atoi(argv[3]); + int M = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + int DiK = atoi(argv[7]); + DP kBT = atof(argv[8]); + DP ommin = atof(argv[9]); + DP ommax = atof(argv[10]); + int Nom = atoi(argv[11]); + DP gwidth = atof(argv[12]); + + stringstream filenameprefix; + Data_File_Name (filenameprefix, whichDSF, Delta, N, M, iKmin, iKmax, kBT, 0, ""); + string prefix = filenameprefix.str(); + + DP normalization = twoPI; + DP denom_sum_K = 1.0/N; + + Write_K_File (N, iKmin, iKmax); + Write_Omega_File (Nom, ommin, ommax); + + DP sumcheck; + sumcheck = Smoothen_RAW_into_SF (prefix, iKmin, iKmax, DiK, ommin, ommax, Nom, gwidth, normalization, denom_sum_K); + + } + + /* + else if (argc == 10) { // fixed_iK + char whichDSF = *argv[1]; + DP Delta = atof(argv[2]); + int N = atoi(argv[3]); + int M = atoi(argv[4]); + int iKneeded = atoi(argv[5]); + DP ommin = atof(argv[6]); + DP ommax = atof(argv[7]); + int Nom = atoi(argv[8]); + DP gwidth = atof(argv[9]); + + bool fixed_iK = true; + + stringstream filenameprefix; + Data_File_Name (filenameprefix, whichDSF, Delta, N, M, fixed_iK, iKneeded, 0.0, 0, ""); + string prefix = filenameprefix.str(); + + DP normalization = twoPI; + int iKmin = iKneeded; + int iKmax = iKneeded; + + cout << "Smoothing: sumcheck = " << Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, gwidth, normalization) << endl; + } + */ + + //else JSCerror("Wrong number of arguments to Smoothen_Heis_DSF executable."); + + return(0); +} diff --git a/src/EXECS/Smoothen_LiebLin_DSF.cc b/src/EXECS/Smoothen_LiebLin_DSF.cc new file mode 100644 index 0000000..6a7e531 --- /dev/null +++ b/src/EXECS/Smoothen_LiebLin_DSF.cc @@ -0,0 +1,111 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Smoothen_LiebLin_DSF.cc + +Purpose: produces .dsf and .ssf files from a .raw file + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 13) { // Print out instructions + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Smoothen_LiebLin_DSF executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers" << endl; + cout << "DP kBT \t\t Temperature" << endl; + cout << "int DiK \t\t\t Window of iK over which DSF is averaged (DiK == 0 means a single iK is used; DiK == 1 means 3 are used (iK-1, iK, iK+1), etc.)" << endl; + cout << "DP ommin" << endl << "DP ommax \t\t Min and max frequencies to cover in smoothened DSF" << endl; + cout << "Nom \t\t\t Number of frequency points used for discretization" << endl; + cout << "DP width \t\t Gaussian width used in smoothing, in units of two-particle level spacing" << endl; + + cout << endl << "EXAMPLE: " << endl << endl; + cout << "Smoothen_LiebLin_DSF d 1.0 100.0 100 0 200 5.0 1 0.0 10.0 500 2.0" << endl << endl; + + } + + else if (argc == 13) { // !fixed_iK + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + int DiK = atoi(argv[8]); + DP ommin = atof(argv[9]); + DP ommax = atof(argv[10]); + int Nom = atoi(argv[11]); + DP width = atof(argv[12]); + + stringstream filenameprefix; + //void Data_File_Name (stringstream& name, char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, DP L2) + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + string prefix = filenameprefix.str(); + + DP normalization = twoPI * L; + DP denom_sum_K = L; + + Write_K_File (L, iKmin, iKmax); + Write_Omega_File (Nom, ommin, ommax); + //cout << "Smoothing: sumcheck = " << Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, width, normalization) << endl; + // We use the scaled width function as default: + + DP sumcheck; + //if (kBT < 0.1) + //sumcheck = Smoothen_RAW_into_SF_LiebLin_Scaled (prefix, L, N, iKmin, iKmax, DiK, ommin, ommax, Nom, width, normalization); + sumcheck = Smoothen_RAW_into_SF (prefix, iKmin, iKmax, DiK, ommin, ommax, Nom, width, normalization, denom_sum_K); + //else sumcheck = Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, width, normalization); + //cout << "Smoothing: sumcheck = " << sumcheck << endl; + } + + /* + else if (argc == 11) { // fixed_iK + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iK_UL = atoi(argv[5]); + int iKneeded = atoi(argv[6]); + DP ommin = atof(argv[7]); + DP ommax = atof(argv[8]); + int Nom = atoi(argv[9]); + DP gwidth = atof(argv[10]); + + //bool fixed_iK = true; + //LiebLin_Bethe_State GroundState (c_int, L, N, iK_UL, 0LL); + stringstream filenameprefix; + //Data_File_Name (filenameprefix, whichDSF, fixed_iK, iKneeded, GroundState, GroundState); + //Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, 0.0); + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iK_UL, iKneeded, iKneeded, 0.0); + string prefix = filenameprefix.str(); + + DP normalization = twoPI * L; + int iKmin = iKneeded; + int iKmax = iKneeded; + + cout << "Smoothing: sumcheck = " << Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, gwidth, normalization) << endl; + } + */ + + //else JSCerror("Wrong number of arguments to Smoothen_LiebLin_DSF executable."); + + return(0); +} diff --git a/src/EXECS/Smoothen_LiebLin_DSF_GeneralState.cc b/src/EXECS/Smoothen_LiebLin_DSF_GeneralState.cc new file mode 100644 index 0000000..1de11f3 --- /dev/null +++ b/src/EXECS/Smoothen_LiebLin_DSF_GeneralState.cc @@ -0,0 +1,120 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Smoothen_LiebLin_DSF_GeneralState.cc + +Purpose: produces .dsf and .ssf files from a .raw file + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 13) { // Print out instructions + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Smoothen_LiebLin_DSF executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "char* defaultScanStatename:\t\t file [].Ix2 contains the quantum numbers defining the AveragingState; used as defaultScanStatename" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers" << endl; + //cout << "DP kBT \t\t Temperature" << endl; + cout << "int DiK \t\t\t Window of iK over which DSF is averaged (DiK == 0 means a single iK is used; DiK == 1 means 3 are used (iK-1, iK, iK+1), etc.)" << endl; + cout << "DP ommin" << endl << "DP ommax \t\t Min and max frequencies to cover in smoothened DSF" << endl; + cout << "Nom \t\t\t Number of frequency points used for discretization" << endl; + cout << "DP width \t\t Gaussian width used in smoothing, in units of two-particle level spacing" << endl; + + cout << endl << "EXAMPLE: " << endl << endl; + cout << "Smoothen_LiebLin_DSF d 1.0 100.0 100 0 200 5.0 1 0.0 10.0 500 2.0" << endl << endl; + + } + + else if (argc == 13) { // !fixed_iK + int n = 1; + char whichDSF = *argv[n++]; + DP c_int = atof(argv[n++]); + DP L = atof(argv[n++]); + int N = atoi(argv[n++]); + char* Ix2filenameprefix = argv[n++]; + int iKmin = atoi(argv[n++]); + int iKmax = atoi(argv[n++]); + //DP kBT = atof(argv[7]); + DP kBT = 0.0; + int DiK = atoi(argv[n++]); + DP ommin = atof(argv[n++]); + DP ommax = atof(argv[n++]); + int Nom = atoi(argv[n++]); + DP width = atof(argv[n++]); + + stringstream filenamestrstream; + filenamestrstream << Ix2filenameprefix; + string defaultScanStatename = filenamestrstream.str(); + + stringstream filenameprefix; + //void Data_File_Name (stringstream& name, char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, DP L2) + //Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, defaultScanStatename); + string prefix = filenameprefix.str(); + + DP normalization = twoPI * L; + DP denom_sum_K = L; + + Write_K_File (L, iKmin, iKmax); + Write_Omega_File (Nom, ommin, ommax); + //cout << "Smoothing: sumcheck = " << Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, width, normalization) << endl; + // We use the scaled width function as default: + + DP sumcheck; + //if (kBT < 0.1) + //sumcheck = Smoothen_RAW_into_SF_LiebLin_Scaled (prefix, L, N, iKmin, iKmax, DiK, ommin, ommax, Nom, width, normalization); + sumcheck = Smoothen_RAW_into_SF (prefix, iKmin, iKmax, DiK, ommin, ommax, Nom, width, normalization, denom_sum_K); + //else sumcheck = Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, width, normalization); + //cout << "Smoothing: sumcheck = " << sumcheck << endl; + } + + /* + else if (argc == 11) { // fixed_iK + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iK_UL = atoi(argv[5]); + int iKneeded = atoi(argv[6]); + DP ommin = atof(argv[7]); + DP ommax = atof(argv[8]); + int Nom = atoi(argv[9]); + DP gwidth = atof(argv[10]); + + //bool fixed_iK = true; + //LiebLin_Bethe_State GroundState (c_int, L, N, iK_UL, 0LL); + stringstream filenameprefix; + //Data_File_Name (filenameprefix, whichDSF, fixed_iK, iKneeded, GroundState, GroundState); + //Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iK_UL, fixed_iK, iKneeded, 0.0); + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iK_UL, iKneeded, iKneeded, 0.0); + string prefix = filenameprefix.str(); + + DP normalization = twoPI * L; + int iKmin = iKneeded; + int iKmax = iKneeded; + + cout << "Smoothing: sumcheck = " << Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, gwidth, normalization) << endl; + } + */ + + //else JSCerror("Wrong number of arguments to Smoothen_LiebLin_DSF executable."); + + return(0); +} diff --git a/src/EXECS/Smoothen_LiebLin_DSF_MosesState.cc b/src/EXECS/Smoothen_LiebLin_DSF_MosesState.cc new file mode 100644 index 0000000..cea0046 --- /dev/null +++ b/src/EXECS/Smoothen_LiebLin_DSF_MosesState.cc @@ -0,0 +1,97 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Smoothen_LiebLin_DSF_MosesState.cc + +Purpose: produces .dsf and .ssf files from a .raw file + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 15) { // Print out instructions + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Smoothen_LiebLin_DSF_MosesState executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "int Nl \t\t\t Number of particles in left Fermi sea (Nr is then N - Nl)" << endl; + cout << "int DIl \t\t shift of left sea as compared to its ground state position" << endl; + cout << "int DIr \t\t shift of right sea as compared to its ground state position" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers" << endl; + //cout << "DP kBT \t\t Temperature" << endl; + cout << "int DiK \t\t\t Window of iK over which DSF is averaged (DiK == 0 means a single iK is used; DiK == 1 means 3 are used (iK-1, iK, iK+1), etc.)" << endl; + cout << "DP ommin" << endl << "DP ommax \t\t Min and max frequencies to cover in smoothened DSF" << endl; + cout << "Nom \t\t\t Number of frequency points used for discretization" << endl; + cout << "DP width \t\t Gaussian width used in smoothing, in units of two-particle level spacing" << endl; + + cout << endl << "EXAMPLE: " << endl << endl; + cout << "Smoothen_LiebLin_DSF_MosesState d 1.0 100.0 100 50 -30 40 0 200 1 0.0 10.0 500 2.0" << endl << endl; + + } + + else if (argc == 15) { // !fixed_iK + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int Nl = atoi(argv[5]); + //int Nr = N - Nl; + int DIl = atoi(argv[6]); + int DIr = atoi(argv[7]); + int iKmin = atoi(argv[8]); + int iKmax = atoi(argv[9]); + //DP kBT = atof(argv[10]); + DP kBT = 0.0; + int DiK = atoi(argv[10]); + DP ommin = atof(argv[11]); + DP ommax = atof(argv[12]); + int Nom = atoi(argv[13]); + DP width = atof(argv[14]); + + // Handy default name: + stringstream defaultScanStatename_strstream; + defaultScanStatename_strstream << "Moses_Nl_" << Nl << "_DIl_" << DIl << "_DIr_" << DIr; + string defaultScanStatename = defaultScanStatename_strstream.str(); + + stringstream filenameprefix; + //void Data_File_Name (stringstream& name, char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, DP L2) + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, defaultScanStatename); + string prefix = filenameprefix.str(); + + + + DP normalization = twoPI * L; + DP denom_sum_K = L; + + Write_K_File (L, iKmin, iKmax); + Write_Omega_File (Nom, ommin, ommax); + //cout << "Smoothing: sumcheck = " << Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, width, normalization) << endl; + // We use the scaled width function as default: + + DP sumcheck; + //if (kBT < 0.1) + //sumcheck = Smoothen_RAW_into_SF_LiebLin_Scaled (prefix, L, N, iKmin, iKmax, DiK, ommin, ommax, Nom, width, normalization); + sumcheck = Smoothen_RAW_into_SF (prefix, iKmin, iKmax, DiK, ommin, ommax, Nom, width, normalization, denom_sum_K); + //else sumcheck = Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, width, normalization); + //cout << "Smoothing: sumcheck = " << sumcheck << endl; + } + + //else JSCerror("Wrong number of arguments to Smoothen_LiebLin_DSF executable."); + + return(0); +} diff --git a/src/EXECS/Smoothen_LiebLin_DSF_Scaled.cc b/src/EXECS/Smoothen_LiebLin_DSF_Scaled.cc new file mode 100644 index 0000000..a2b08ea --- /dev/null +++ b/src/EXECS/Smoothen_LiebLin_DSF_Scaled.cc @@ -0,0 +1,82 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Smoothen_LiebLin_DSF_Scaled.cc + +Purpose: produces .dsfs and .ssf files from a .raw file + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 13) { // Print out instructions + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Smoothen_LiebLin_DSF_Scaled executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers" << endl; + cout << "DP kBT \t\t Temperature" << endl; + cout << "int DiK \t\t\t Window of iK over which DSF is averaged (DiK == 0 means a single iK is used; DiK == 1 means 3 are used (iK-1, iK, iK+1), etc.)" << endl; + cout << "DP ommin" << endl << "DP ommax \t\t Min and max frequencies to cover in smoothened DSF" << endl; + cout << "Nom \t\t\t Number of frequency points used for discretization" << endl; + cout << "DP width \t\t Gaussian width used in smoothing, in units of two-particle level spacing" << endl; + + cout << endl << "EXAMPLE: " << endl << endl; + cout << "Smoothen_LiebLin_DSF d 1.0 100.0 100 0 200 5.0 1 0.0 10.0 500 2.0" << endl << endl; + + } + + else if (argc == 13) { // !fixed_iK + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + int DiK = atoi(argv[8]); + DP ommin = atof(argv[9]); + DP ommax = atof(argv[10]); + int Nom = atoi(argv[11]); + DP width = atof(argv[12]); + + stringstream filenameprefix; + //void Data_File_Name (stringstream& name, char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, DP L2) + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + string prefix = filenameprefix.str(); + + DP normalization = twoPI * L; + //DP denom_sum_K = L; + + Write_K_File (L, iKmin, iKmax); + Write_Omega_File (Nom, ommin, ommax); + //cout << "Smoothing: sumcheck = " << Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, width, normalization) << endl; + // We use the scaled width function as default: + + DP sumcheck; + //if (kBT < 0.1) + sumcheck = Smoothen_RAW_into_SF_LiebLin_Scaled (prefix, L, N, iKmin, iKmax, DiK, ommin, ommax, Nom, width, normalization); + //sumcheck = Smoothen_RAW_into_SF (prefix, iKmin, iKmax, DiK, ommin, ommax, Nom, width, normalization, denom_sum_K); + //else sumcheck = Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, width, normalization); + //cout << "Smoothing: sumcheck = " << sumcheck << endl; + } + + //else JSCerror("Wrong number of arguments to Smoothen_LiebLin_DSF executable."); + + return(0); +} diff --git a/src/EXECS/Smoothen_LiebLin_DSF_over_Ensemble.cc b/src/EXECS/Smoothen_LiebLin_DSF_over_Ensemble.cc new file mode 100644 index 0000000..2d0c09d --- /dev/null +++ b/src/EXECS/Smoothen_LiebLin_DSF_over_Ensemble.cc @@ -0,0 +1,103 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Smoothen_LiebLin_DSF_over_Ensemble.cc + +Purpose: produces .dsf and .ssf files from an ensemble of .raw files + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 13) { // Print out instructions + + cout << endl << "Welcome to ABACUS++\t(copyright J.-S. Caux)." << endl; + cout << endl << "Usage of Smoothen_LiebLin_DSF_over_Ensemble executable: " << endl; + cout << endl << "Provide the following arguments:" << endl << endl; + cout << "char whichDSF \t\t Which structure factor should be calculated ? Options are: d for rho rho, g for psi psi{dagger}, o for psi{dagger} psi" << endl; + cout << "DP c_int \t\t Value of the interaction parameter" << endl; + cout << "DP L \t\t\t Length of the system" << endl; + cout << "int N \t\t\t Number of particles" << endl; + cout << "int iKmin" << endl << "int iKmax \t\t Min and max momentum integers" << endl; + cout << "DP kBT \t\t Temperature" << endl; + //cout << "int nstates \t\t\t Number of states considered in the ensemble" << endl; + cout << "int DiK \t\t\t Window of iK over which DSF is averaged (DiK == 0 means a single iK is used; DiK == 1 means 3 are used (iK-1, iK, iK+1), etc.)" << endl; + cout << "DP ommin" << endl << "DP ommax \t\t Min and max frequencies to cover in smoothened DSF" << endl; + cout << "Nom \t\t\t Number of frequency points used for discretization" << endl; + cout << "DP width \t\t Gaussian width used in smoothing, in units of two-particle level spacing" << endl; + + cout << endl << "EXAMPLE: " << endl << endl; + cout << "Smoothen_LiebLin_DSF_over_Ensemble d 1.0 100.0 100 0 200 5.0 1 0.0 10.0 500 2.0" << endl << endl; + + } + + else if (argc == 13) { + char whichDSF = *argv[1]; + DP c_int = atof(argv[2]); + DP L = atof(argv[3]); + int N = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP kBT = atof(argv[7]); + //int nstates_req = atoi(argv[8]); + int DiK = atoi(argv[8]); + DP ommin = atof(argv[9]); + DP ommax = atof(argv[10]); + int Nom = atoi(argv[11]); + DP width = atof(argv[12]); + + stringstream filenameprefix; + //void Data_File_Name (stringstream& name, char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, DP L2) + Data_File_Name (filenameprefix, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + string prefix = filenameprefix.str(); + + DP normalization = twoPI * L; + DP denom_sum_K = L; + + Write_K_File (L, iKmin, iKmax); + Write_Omega_File (Nom, ommin, ommax); + + // Read the weights from the ensembles file: + LiebLin_Diagonal_State_Ensemble ensemble; + + stringstream ensfilestrstream; + //ensfilestrstream << "LiebLin_c_int_" << c_int << "_L_" << L << "_N_" << N << "_kBT_" << kBT << "_ns_" << nstates_req << ".ens"; + ensfilestrstream << "LiebLin_c_int_" << c_int << "_L_" << L << "_N_" << N << "_kBT_" << kBT << ".ens"; + string ensfilestr = ensfilestrstream.str(); + const char* ensfile_Cstr = ensfilestr.c_str(); + + ensemble.Load(c_int, L, N, ensfile_Cstr); + + // Define the raw file names + Vect rawfilename(ensemble.nstates); + + for (int ns = 0; ns < ensemble.nstates; ++ns) { + // Define the raw input file name: + stringstream filenameprefix; + //Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, kBT, ensemble.state[ns], ensemble.state[ns], ensemble.state[ns].label); + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, 0.0, ensemble.state[ns], ensemble.state[ns], ensemble.state[ns].label); + string prefix = filenameprefix.str(); + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + //RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + rawfilename[ns] = RAW_stringstream.str(); + } + + Smoothen_RAW_into_SF (prefix, rawfilename, ensemble.weight, iKmin, iKmax, DiK, + ommin, ommax, Nom, width, normalization, denom_sum_K); + + } + + return(0); +} diff --git a/src/EXECS/Smoothen_ODSLF_DSF.cc b/src/EXECS/Smoothen_ODSLF_DSF.cc new file mode 100644 index 0000000..a343dde --- /dev/null +++ b/src/EXECS/Smoothen_ODSLF_DSF.cc @@ -0,0 +1,82 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Smoothen_ODSLF_DSF.cc + +Purpose: produces .dsf and .ssf files from a .raw file + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main(int argc, char* argv[]) +{ + if (argc != 10 && argc != 11) { // Print out instructions + //if (strcmp(argv[1],"help") == 0) { // Output some instructions + cout << "Usage of Smoothen_ODSLF_DSF executable: " << endl << endl; + cout << "Provide arguments using one of the following options:" << endl << endl; + cout << "1) (for general momenta) whichDSF Delta N M iKmin iKmax ommin ommax Nom gwidth" << endl << endl; + cout << "2) (for fixed momentum) whichDSF Delta N M iKneeded ommin ommax Nom gwidth" << endl << endl; + //else JSCerror("Incomprehensible arguments in Smoothen_ODSLF_DSF executable."); + } + + else if (argc == 11) { // !fixed_iK + char whichDSF = *argv[1]; + DP Delta = atof(argv[2]); + int N = atoi(argv[3]); + int M = atoi(argv[4]); + int iKmin = atoi(argv[5]); + int iKmax = atoi(argv[6]); + DP ommin = atof(argv[7]); + DP ommax = atof(argv[8]); + int Nom = atoi(argv[9]); + DP gwidth = atof(argv[10]); + + stringstream filenameprefix; + ODSLF_Data_File_Name (filenameprefix, whichDSF, Delta, N, M, iKmin, iKmax, 0.0, 0); + string prefix = filenameprefix.str(); + + DP normalization = twoPI; + + cout << "Smoothing: sumcheck = " << Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, gwidth, normalization) << endl; + + Write_K_File (N, iKmin, iKmax); + Write_Omega_File (Nom, ommin, ommax); + + } + + else if (argc == 10) { // fixed_iK + char whichDSF = *argv[1]; + DP Delta = atof(argv[2]); + int N = atoi(argv[3]); + int M = atoi(argv[4]); + int iKneeded = atoi(argv[5]); + DP ommin = atof(argv[6]); + DP ommax = atof(argv[7]); + int Nom = atoi(argv[8]); + DP gwidth = atof(argv[9]); + + bool fixed_iK = true; + + stringstream filenameprefix; + Data_File_Name (filenameprefix, whichDSF, Delta, N, M, fixed_iK, iKneeded, 0.0, 0); + string prefix = filenameprefix.str(); + + DP normalization = twoPI; + int iKmin = iKneeded; + int iKmax = iKneeded; + + cout << "Smoothing: sumcheck = " << Smoothen_RAW_into_SF (prefix, iKmin, iKmax, ommin, ommax, Nom, gwidth, normalization) << endl; + } + + else JSCerror("Wrong number of arguments to Smoothen_Heis_DSF executable."); +} diff --git a/src/EXECS/XXZ_gpd_StagSz_h0.cc b/src/EXECS/XXZ_gpd_StagSz_h0.cc new file mode 100644 index 0000000..7ea767e --- /dev/null +++ b/src/EXECS/XXZ_gpd_StagSz_h0.cc @@ -0,0 +1,112 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: ABACUS++G_2_testing.cc + +Purpose: testing of ABACUS++2 + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +int main( int argc, char* argv[]) +{ + if (!(argc == 3 || argc == 5)) { // provide some info + cout << endl << "This code computes the (1/N) (-1)^j S^z_j on-site staggered magnetization for XXZ_gpd in zero field." << endl; + cout << "First option: provide two arguments: anisotropy Delta (> 1) and system size N (even)." << endl; + cout << "Second option: provide five arguments: system size N (even), Delta min, Delta max, NDelta." << endl; + cout << "The output is Delta, N, stag mag, energy gap." << endl; + JSCerror(""); + } + else if (argc == 3) { + DP Delta = atof(argv[1]); + if (Delta <= 1.0) JSCerror("Provide Delta > 1."); + int N = atoi(argv[2]); + if (N % 2) JSCerror("Provide an even Delta."); + int M = N/2; + + // Define the chain: J, Delta, h, Nsites + Heis_Chain chain(1.0, Delta, 0.0, N); + + Heis_Base gbase(chain, M); + + XXZ_gpd_Bethe_State gstate(chain, gbase); + gstate.Compute_All(true); + + XXZ_gpd_Bethe_State estate(chain, gbase); + estate.Ix2[0][0] = M+1; // umklapp excitation + estate.Compute_All(true); + + stringstream basestrstream; + basestrstream << M-2 << "x1"; + string basestr = basestrstream.str(); + Heis_Base basegap(chain, basestr); + XXZ_gpd_Bethe_State estategap(chain, basegap); + estategap.Compute_All(true); + + if (!gstate.conv) JSCerror("Ground state did not converge."); + if (!estate.conv) JSCerror("Umklapp state did not converge."); + if (!estategap.conv) JSCerror("Gap state did not converge."); + + cout << Delta << "\t" << N << "\t" << setprecision(12) << exp(real(ln_Sz_ME (gstate, estate)))/sqrt(N) << "\t" << estategap.E - gstate.E << endl; + + } + + else if (argc == 5) { // Do a scan in Delta + + int N = atoi(argv[1]); + if (N % 2) JSCerror("Provide an even Delta."); + int M = N/2; + + DP Deltamin = atof(argv[2]); + if (Deltamin <= 1.0) JSCerror("Provide Deltamin > 1."); + + DP Deltamax = atof(argv[3]); + if (Deltamin <= 1.0) JSCerror("Provide Deltamax > Deltamin."); + + int NDelta = atoi(argv[4]); + + for (int iDelta = 0; iDelta <= NDelta; ++iDelta) { + + DP Delta = (Deltamin * (NDelta -iDelta) + Deltamax * iDelta)/NDelta; + + // Define the chain: J, Delta, h, Nsites + Heis_Chain chain(1.0, Delta, 0.0, N); + + Heis_Base gbase(chain, M); + + XXZ_gpd_Bethe_State gstate(chain, gbase); + gstate.Compute_All(true); + + XXZ_gpd_Bethe_State estate(chain, gbase); + estate.Ix2[0][0] = M+1; // umklapp excitation + estate.Compute_All(true); + + stringstream basestrstream; + basestrstream << M-2 << "x1"; + string basestr = basestrstream.str(); + Heis_Base basegap(chain, basestr); + XXZ_gpd_Bethe_State estategap(chain, basegap); + estategap.Compute_All(true); + + if (!gstate.conv) JSCerror("Ground state did not converge."); + if (!estate.conv) JSCerror("Umklapp state did not converge."); + if (!estategap.conv) JSCerror("Gap state did not converge."); + + cout << Delta << "\t" << N << "\t" << setprecision(12) << exp(real(ln_Sz_ME (gstate, estate)))/sqrt(N) << "\t" << estategap.E - gstate.E << endl; + + } + } + + return(0); +} diff --git a/src/FITTING/covsrt.cc b/src/FITTING/covsrt.cc new file mode 100644 index 0000000..598869a --- /dev/null +++ b/src/FITTING/covsrt.cc @@ -0,0 +1,41 @@ +/********************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) 2007. + +----------------------------------------------------------- + +File: mrqmin.cc + +Purpose: Nonlinear fitting + +Last modified: 14/08/07 + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + void covsrt (SQMat_DP& covar, Vect& ia, const int mfit) + { + int i, j, k; + + int ma = ia.size(); + + for (i = mfit; i < ma; i++) + for (j = 0; j < i+1; j++) covar[i][j] = covar[j][i] = 0.0; + k = mfit - 1; + for (j = ma - 1; j >= 0; j--) { + if (ia[j]) { + for (i = 0; i < ma; i++) SWAP(covar[i][k], covar[i][j]); + for (i = 0; i < ma; i++) SWAP(covar[k][i], covar[j][i]); + k--; + } + } + } + +} diff --git a/src/FITTING/lin_reg.cc b/src/FITTING/lin_reg.cc new file mode 100644 index 0000000..8ea33b4 --- /dev/null +++ b/src/FITTING/lin_reg.cc @@ -0,0 +1,73 @@ +/********************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) 2006. + +----------------------------------------------------------- + +File: lin_reg.cc + +Purpose: Linear regression + +Last modified: 11/05/07 + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + void lin_reg (Vect_DP x, Vect_DP y, Vect_DP sigma, DP& a, DP& b, DP& chisq) + { + // Performs simple linear regression on data + + int Npts = x.size(); + + DP S = 0.0; + DP Sx = 0.0; + DP Sy = 0.0; + DP Stt = 0.0; + Vect_DP t(0.0, Npts); + + for (int i = 0; i < Npts; ++i) { + S += 1.0/(sigma[i] * sigma[i]); + Sx += x[i]/(sigma[i] * sigma[i]); + Sy += y[i]/(sigma[i] * sigma[i]); + } + + for (int i = 0; i < Npts; ++i) { + t[i] = (x[i] - Sx/S)/sigma[i]; + Stt += t[i] * t[i]; + } + + a = 0.0; + b = 0.0; + for (int i = 0; i < Npts; ++i) { + b += t[i] * y[i]/sigma[i]; + } + b /= Stt; + a = (Sy - Sx * b)/S; + + chisq = 0.0; + for (int i = 0; i < Npts; ++i) { + chisq += (y[i] - a - b * x[i]) * (y[i] - a - b * x[i]) / (sigma[i] * sigma[i]); + } + + return; + } + + void lin_reg (Vect_DP x, Vect_DP y, DP& a, DP& b, DP& chisq) + { + // Assumes all sigma == 1 + + Vect_DP sigma(1.0, x.size()); + + lin_reg (x, y, sigma, a, b, chisq); + + return; + } + +} diff --git a/src/FITTING/mrq.cc b/src/FITTING/mrq.cc new file mode 100644 index 0000000..2bf62f2 --- /dev/null +++ b/src/FITTING/mrq.cc @@ -0,0 +1,127 @@ +/********************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) 2007. + +----------------------------------------------------------- + +File: mrqmin.cc + +Purpose: Nonlinear fitting + +Last modified: 14/08/07 + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + void mrqmin (Vect_DP& x, Vect_DP& y, Vect_DP& sig, Vect_DP& a, + Vect& ia, SQMat_DP& covar, SQMat_DP& alpha, DP& chisq, + void funcs(const DP, Vect_DP&, DP&, Vect_DP&), DP& alambda) + { + // Levenberg-Marquardt method. See NRC++ p.691. + + static int mfit; + static DP ochisq; + int j, k, l; + + int ma = a.size(); + static SQMat_DP oneda (ma); + static Vect_DP atry(ma); + static Vect_DP beta(ma); + static Vect_DP da(ma); + + if (alambda < 0.0) { + mfit = 0; + for (j = 0; j < ma; j++) + if (ia[j]) mfit++; + alambda = 0.001; + mrqcof (x, y, sig, a, ia, alpha, beta, chisq, funcs); + ochisq = chisq; + for (j = 0; j < ma; j++) atry[j] = a[j]; + } + + SQMat_DP temp (mfit); + + for (j = 0; j < mfit; j++) { + for (k = 0; k < mfit; k++) covar[j][k] = alpha[j][k]; + covar[j][j] = alpha[j][j] * (1.0 + alambda); + for (k = 0; k < mfit; k++) temp[j][k] = covar[j][k]; + oneda[j][0] = beta[j]; + } + + gaussj (temp, oneda); + for (j = 0; j < mfit; j++) { + for (k = 0; k < mfit; k++) covar[j][k] = temp[j][k]; + da[j] = oneda[j][0]; + } + + if (alambda == 0.0) { + covsrt (covar, ia, mfit); + covsrt (alpha, ia, mfit); + return; + } + + for (j = 0, l = 0; l < ma; l++) + if (ia[l]) atry[l] = a[l] + da[j++]; + + mrqcof (x, y, sig, atry, ia, covar, da, chisq, funcs); + + if (chisq < ochisq) { + alambda *= 0.1; + ochisq = chisq; + for (j = 0; j < mfit; j++) { + for (k = 0; k < mfit; k++) alpha[j][k] = covar[j][k]; + beta[j] = da[j]; + } + for (l = 0; l < ma; l++) a[l] = atry[l]; + } else { + alambda *= 10.0; + chisq = ochisq; + } + } + + void mrqcof (Vect_DP& x, Vect_DP& y, Vect_DP& sig, Vect_DP& a, + Vect& ia, SQMat_DP& alpha, Vect_DP& beta, DP& chisq, + void funcs (const DP, Vect_DP&, DP&, Vect_DP&)) + { + int i, j, k, l, m, mfit = 0; + DP ymod, wt, sig2i, dy; + + int ndata = x.size(); + int ma = a.size(); + + Vect_DP dyda (ma); + + for (j = 0; j < ma; j++) + if (ia[j]) mfit++; + for (j = 0; j < mfit; j++) { + for (k = 0; k <= j; k++) alpha[j][k] = 0.0; + beta[j] = 0.0; + } + + chisq = 0.0; + for (i = 0; i < ndata; i++) { + funcs (x[i], a, ymod, dyda); + sig2i = 1.0/(sig[i] * sig[i]); + dy = y[i] - ymod; + for (j = 0, l = 0; l < ma; l++) { + if (ia[l]) { + wt = dyda[l] * sig2i; + for (k = 0, m = 0; m < l+1; m++) + if (ia[m]) alpha[j][k++] += wt * dyda[m]; + beta[j++] += dy * wt; + } + } + chisq += dy * dy * sig2i; + } + for (j = 1; j < mfit; j++) + for (k = 0; k < j; k++) alpha[k][j] = alpha[j][k]; + } + +} diff --git a/src/FITTING/polint.cc b/src/FITTING/polint.cc new file mode 100644 index 0000000..e6f1cc1 --- /dev/null +++ b/src/FITTING/polint.cc @@ -0,0 +1,35 @@ +#include "JSC.h" +using namespace std; + +void JSC::polint(Vect_DP& xa, Vect_DP& ya, const DP x, DP& y, DP& dy) +{ + // Polynomial interpolation/extrapolation, NR page 113. + + int i, m, ns=0; + DP den, dif, dift, ho, hp, w; + + int n = xa.size(); + Vect_DP c(n), d(n); + dif = fabs(x-xa[0]); + for (i = 0; i < n; i++) { + if ((dift = fabs(x-xa[i])) < dif) { + ns = i; + dif = dift; + } + c[i] = ya[i]; + d[i] = ya[i]; + } + y = ya[ns--]; + for (m = 1; m < n; m++) { + for (i = 0; i < n-m; i++) { + ho = xa[i] - x; + hp = xa[i+m] - x; + w = c[i+1] - d[i]; + if ((den = ho-hp) == 0.0) JSCerror("Error in routine polint."); + den = w/den; + d[i] = hp * den; + c[i] = ho * den; + } + y += (dy = (2 * (ns + 1) < (n-m) ? c[ns + 1] : d[ns--])); + } +} diff --git a/src/FITTING/polint_cx.cc b/src/FITTING/polint_cx.cc new file mode 100644 index 0000000..ae6edb8 --- /dev/null +++ b/src/FITTING/polint_cx.cc @@ -0,0 +1,36 @@ +#include "JSC.h" +using namespace std; + +void JSC::polint(Vect_CX& xa, Vect_CX& ya, const complex x, complex& y, complex& dy) +{ + // Polynomial interpolation/extrapolation, NR page 113. + + int i, m, ns=0; + DP dif, dift; + complex den, ho, hp, w; + + int n = xa.size(); + Vect_CX c(n), d(n); + dif = abs(x-xa[0]); + for (i = 0; i < n; i++) { + if ((dift = abs(x-xa[i])) < dif) { + ns = i; + dif = dift; + } + c[i] = ya[i]; + d[i] = ya[i]; + } + y = ya[ns--]; + for (m = 1; m < n; m++) { + for (i = 0; i < n-m; i++) { + ho = xa[i] - x; + hp = xa[i+m] - x; + w = c[i+1] - d[i]; + if ((den = ho-hp) == 0.0) JSCerror("Error in routine polint_cx."); + den = w/den; + d[i] = hp * den; + c[i] = ho * den; + } + y += (dy = (2 * (ns + 1) < (n-m) ? c[ns + 1] : d[ns--])); + } +} diff --git a/src/HEIS/Heis.cc b/src/HEIS/Heis.cc new file mode 100644 index 0000000..cfa5519 --- /dev/null +++ b/src/HEIS/Heis.cc @@ -0,0 +1,1717 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/HEIS/Heis.cc + +Purpose: defines functions in all HEIS classes. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + //*************************************************************************************************** + + // Function definitions: class Heis_Chain + + Heis_Chain::Heis_Chain() + : J(0.0), Delta(0.0), anis(0.0), hz(0.0), Nsites(0), Nstrings(0), Str_L(0), par(0), + si_n_anis_over_2(new DP[1]), co_n_anis_over_2(new DP[1]), ta_n_anis_over_2(new DP[1]), prec(ITER_REQ_PREC) {} + + Heis_Chain::Heis_Chain (DP JJ, DP DD, DP hh, int NN) + : J(JJ), Delta(DD), anis(0.0), hz(hh), Nsites(NN), Nstrings(MAXSTRINGS), Str_L(new int[MAXSTRINGS]), par(new int[MAXSTRINGS]), + si_n_anis_over_2(new DP[10*MAXSTRINGS]), co_n_anis_over_2(new DP[10*MAXSTRINGS]), ta_n_anis_over_2(new DP[10*MAXSTRINGS]), prec(ITER_REQ_PREC) + { + // We restrict to even chains everywhere + + if (NN % 2) JSCerror("Please use an even-length chain."); + + if ((Delta > 0.0) && (Delta < 1.0)) { + + // gapless XXZ case + + anis = acos(DD); + + // Set the Str_L and par vectors: + + DP gammaoverpi = acos(DD)/PI; + //cout << "gammaoverpi = " << gammaoverpi << endl; + + Vect Nu(MAXSTRINGS); + + DP gammaoverpi_eff = gammaoverpi; + DP gammaoverpi_reached = 0.0; + + // First, define the series Nu[i], like in TakahashiBOOK eqn 9.35 + // CAREFUL: the labelling is different, Nu[0] is \nu_1, etc. + int l = 0; + int ml_temp = 0; // checks the sum of all Nu's + + while (fabs(gammaoverpi - gammaoverpi_reached) > 1.0e-8){ + + Nu[l] = int(1.0/gammaoverpi_eff); + ml_temp += Nu[l]; + if (Nu[l] == 0) gammaoverpi_eff = 0.0; + else gammaoverpi_eff = 1.0/gammaoverpi_eff - DP(Nu[l]); + + // compute gammaoverpi_reached: + gammaoverpi_reached = Nu[l]; + for (int p = 0; p < l; ++p) gammaoverpi_reached = Nu[l - p - 1] + 1.0/gammaoverpi_reached; + gammaoverpi_reached = 1.0/gammaoverpi_reached; + + //cout << "gammaoverpi_reached = " << gammaoverpi_reached << "\tdiff = " << fabs(gammaoverpi - gammaoverpi_reached) << endl; + //cout << "ml_temp = " << ml_temp << "\tMAXSTRINGS = " << MAXSTRINGS << endl; + + l++; + + if (ml_temp > MAXSTRINGS) break; // we defined Str_L and par as arrays of at most MAXSTRINGS elements, so we cut off here... + + } + //cout << "l = " << l << endl; + + // Check: make sure the last Nu is greater than one: if one, add 1 to previous Nu + + if (l > 1) { + if (Nu[l-1] == 1) { + Nu[l-1] = 0; + Nu[l-2] += 1; + l -= 1; + } + } + + // Length of continued fraction is l-1, which is denoted l in Takahashi + + // Second, define the series y[i] and m[i] as in TakahashiBOOK eqn 9.36 + // y_{-1} is treated separately, and here y[i] = y_i, m[i] = m_i, i = 0, ..., l + Vect y(0, l+1); + Vect m(0, l+1); + y[0] = 1; + m[0] = 0; + y[1] = Nu[0]; + m[1] = Nu[0]; + for (int i = 2; i <= l; ++i) { + y[i] = y[i - 2] + Nu[i-1]*y[i-1]; + m[i] = 0; + for (int k = 0; k < i; ++k) m[i] += Nu[k]; + } + // Now determine the lengths and parity, following TakahashiBOOK eqn 9.37 + // Nstrings = JSC::min(m[l] + 1, MAXSTRINGS); // number of different strings that are possible for this chain + // Always set to MAXSTRINGS + + Str_L[0] = 1; + par[0] = 1; + Str_L[1] = 1; + par[1] = -1; + + int max_j = 0; + + for (int i = 0; i < l; ++i) { + for (int j = JSC::max(1, m[i]) + 1; j < JSC::min(m[i+1] + 1, MAXSTRINGS); ++j) { + if (i == 0) Str_L[j] = (j - m[0])* y[0]; + else if (i >= 1) Str_L[j] = y[i-1] + (j - m[i])*y[i]; + par[j] = (int(floor((Str_L[j] - 1)*gammaoverpi)) % 2) ? -1 : 1; + max_j = j; + } + } + + // Set the rest of Str_L and par vector elements to zero + + for (int i = max_j + 1; i < MAXSTRINGS; ++i) { + Str_L[i] = 0; + par[i] = 0; + } + + // Set the sin, cos and tan_n_zeta_over_2 vectors: + + for (int i = 0; i < 10*MAXSTRINGS; ++i) si_n_anis_over_2[i] = sin(i * anis/2.0); + for (int i = 0; i < 10*MAXSTRINGS; ++i) co_n_anis_over_2[i] = cos(i * anis/2.0); + for (int i = 0; i < 10*MAXSTRINGS; ++i) ta_n_anis_over_2[i] = tan(i * anis/2.0); + + } // if XXZ gapless + + else if (Delta == 1.0) { + + // Isotropic antiferromagnet + + anis = 0.0; + + // Set Str_L and par: + + for (int i = 0; i < MAXSTRINGS; ++i) { + Str_L[i] = i + 1; + par[i] = 1; + } + + } // if XXX AFM + + else if (Delta > 1.0) { + + // Gapped antiferromagnet + + anis = acosh(DD); + + // Set the Str_L and par vectors: + + for (int i = 0; i < MAXSTRINGS; ++i) { + Str_L[i] = i + 1; + par[i] = 1; + } + + // Set the sinh, cosh and tanh_n_eta_over_2 vectors: + + for (int i = 0; i < 10*MAXSTRINGS; ++i) si_n_anis_over_2[i] = sinh(i * anis/2.0); + for (int i = 0; i < 10*MAXSTRINGS; ++i) co_n_anis_over_2[i] = cosh(i * anis/2.0); + for (int i = 0; i < 10*MAXSTRINGS; ++i) ta_n_anis_over_2[i] = tanh(i * anis/2.0); + + } // if XXZ_gpd + + } + + Heis_Chain::Heis_Chain (const Heis_Chain& RefChain) // copy constructor + { + J = RefChain.J; + Delta = RefChain.Delta; + anis = RefChain.anis; + hz = RefChain.hz; + Nsites = RefChain.Nsites; + Nstrings = RefChain.Nstrings; + Str_L = new int[RefChain.Nstrings]; + for (int i = 0; i < RefChain.Nstrings; ++i) Str_L[i] = RefChain.Str_L[i]; + par = new int[RefChain.Nstrings]; + for (int i = 0; i < RefChain.Nstrings; ++i) par[i] = RefChain.par[i]; + si_n_anis_over_2 = new DP[10*MAXSTRINGS]; + for (int i = 0; i < 10*MAXSTRINGS; ++i) si_n_anis_over_2[i] = RefChain.si_n_anis_over_2[i]; + co_n_anis_over_2 = new DP[10*MAXSTRINGS]; + for (int i = 0; i < 10*MAXSTRINGS; ++i) co_n_anis_over_2[i] = RefChain.co_n_anis_over_2[i]; + ta_n_anis_over_2 = new DP[10*MAXSTRINGS]; + for (int i = 0; i < 10*MAXSTRINGS; ++i) ta_n_anis_over_2[i] = RefChain.ta_n_anis_over_2[i]; + prec = RefChain.prec; + } + + Heis_Chain& Heis_Chain::operator= (const Heis_Chain& RefChain) // assignment operator + { + if (this != &RefChain) { + J = RefChain.J; + Delta = RefChain.Delta; + anis = RefChain.anis; + hz = RefChain.hz; + Nsites = RefChain.Nsites; + Nstrings = RefChain.Nstrings; + if (Str_L != 0) delete[] Str_L; + Str_L = new int[RefChain.Nstrings]; + for (int i = 0; i < RefChain.Nstrings; ++i) Str_L[i] = RefChain.Str_L[i]; + if (par != 0) delete[] par; + par = new int[RefChain.Nstrings]; + for (int i = 0; i < RefChain.Nstrings; ++i) par[i] = RefChain.par[i]; + if (si_n_anis_over_2 != 0) delete[] si_n_anis_over_2; + si_n_anis_over_2 = new DP[10*MAXSTRINGS]; + for (int i = 0; i < 10*MAXSTRINGS; ++i) si_n_anis_over_2[i] = RefChain.si_n_anis_over_2[i]; + if (co_n_anis_over_2 != 0) delete[] co_n_anis_over_2; + co_n_anis_over_2 = new DP[10*MAXSTRINGS]; + for (int i = 0; i < 10*MAXSTRINGS; ++i) co_n_anis_over_2[i] = RefChain.co_n_anis_over_2[i]; + if (ta_n_anis_over_2 != 0) delete[] ta_n_anis_over_2; + ta_n_anis_over_2 = new DP[10*MAXSTRINGS]; + for (int i = 0; i < 10*MAXSTRINGS; ++i) ta_n_anis_over_2[i] = RefChain.ta_n_anis_over_2[i]; + prec = RefChain.prec; + + } + return *this; + } + + bool Heis_Chain::operator== (const Heis_Chain& RefChain) + { + bool answer = true; + + if ((J != RefChain.J) || (Nsites != RefChain.Nsites) || (Delta != RefChain.Delta)) answer = false; + + return(answer); + } + + bool Heis_Chain::operator!= (const Heis_Chain& RefChain) + { + bool answer = false; + + if ((J != RefChain.J) || (Nsites != RefChain.Nsites) || (Delta != RefChain.Delta)) answer = true; + + return(answer); + } + + Heis_Chain::~Heis_Chain() + { + if (Str_L != 0) delete[] Str_L; + if (par != 0) delete[] par; + if (si_n_anis_over_2 != 0) delete[] si_n_anis_over_2; + if (co_n_anis_over_2 != 0) delete[] co_n_anis_over_2; + if (ta_n_anis_over_2 != 0) delete[] ta_n_anis_over_2; + } + + /* Deactivated in ++G_8 + void Heis_Chain::Scan_for_Possible_Bases (int Mdown_remaining, Vect& possible_base_label, int& nfound, int nexc_max_used, + int base_level_to_scan, Vect& Nrapidities) + { + if (Mdown_remaining < 0) { JSCerror("Scan_for_Possible_Bases: shouldn't be here..."); } // reached inconsistent point + + //cout << "Mdown_remaining " << Mdown_remaining << "\t" << possible_base_id << "\tnfound " << nfound + // << "\tnexc_max_used " << nexc_max_used << "\tbase_level_to_scan " << base_level_to_scan << "\tNrap " << Nrapidities << endl; + + if (base_level_to_scan == 0) { + Nrapidities[0] = Mdown_remaining; + + // Set label: + stringstream M0out; + M0out << Nrapidities[0]; + possible_base_label[nfound] = M0out.str(); + for (int itype = 1; itype < Nrapidities.size(); ++itype) + if (Nrapidities[itype] > 0) { + possible_base_label[nfound] += TYPESEP; + stringstream typeout; + typeout << itype; + possible_base_label[nfound] += typeout.str(); + possible_base_label[nfound] += EXCSEP; + stringstream Mout; + Mout << Nrapidities[itype]; + possible_base_label[nfound] += Mout.str(); + } + nfound++; + } + + else { + // Remove all higher strings: + //Nrapidities[base_level_to_scan] = 0; + //Scan_for_Possible_Bases (Mdown_remaining, possible_base_id, nfound, nexc_max_used, base_level_to_scan - 1, Nrapidities); + + for (int i = 0; i <= (Str_L[base_level_to_scan] == 0 ? 0 : nexc_max_used/Str_L[base_level_to_scan]); ++i) { + Nrapidities[base_level_to_scan] = i; + Scan_for_Possible_Bases (Mdown_remaining - i*Str_L[base_level_to_scan], possible_base_label, nfound, + nexc_max_used - i*Str_L[base_level_to_scan], base_level_to_scan - 1, Nrapidities); + } + + } + } + + + Vect Heis_Chain::Possible_Bases (int Mdown) // returns a vector of possible bases + { + // We partition Mdown into up to NEXC_MAX_HEIS excitations + + int nexc_max_used = JSC::min(NEXC_MAX_HEIS, 2*(Mdown/2)); // since each inner sector can contain at most N/2 holes. + + Vect possible_base_label (1000); + int nfound = 0; + Vect Nrapidities (0, Nstrings); + + //cout << "In Possible_Bases: start scan for Mdown = " << Mdown << endl; + + (*this).Scan_for_Possible_Bases (Mdown, possible_base_label, nfound, nexc_max_used, Nstrings - 1, Nrapidities); + + // Copy results into a clean vector: + Vect possible_base_label_found (nfound); + for (int i = 0; i < nfound; ++i) possible_base_label_found[i] = possible_base_label[i]; + + //cout << "In Possible_Bases: possible_base_label_found = " << possible_base_label_found << endl; + + return(possible_base_label_found); + } + */ + + //*************************************************************************************************** + + // Function definitions: class Heis_Base + + Heis_Base::Heis_Base () : Mdown(0), Nrap(Vect()), Nraptot(0), + Ix2_infty(Vect()), Ix2_min(Vect()), Ix2_max(Vect()), dimH(0.0), baselabel("") {} + + Heis_Base::Heis_Base (const Heis_Base& RefBase) // copy constructor + : Mdown(RefBase.Mdown), Nrap(Vect(RefBase.Nrap.size())), Nraptot(RefBase.Nraptot), + Ix2_infty(Vect(RefBase.Nrap.size())), Ix2_min(Vect(RefBase.Nrap.size())), Ix2_max(Vect(RefBase.Nrap.size())), + baselabel(RefBase.baselabel) + { + for (int i = 0; i < Nrap.size(); ++i) { + Nrap[i] = RefBase.Nrap[i]; + Ix2_infty[i] = RefBase.Ix2_infty[i]; + Ix2_min[i] = RefBase.Ix2_min[i]; + Ix2_max[i] = RefBase.Ix2_max[i]; + dimH = RefBase.dimH; + } + } + + Heis_Base::Heis_Base (const Heis_Chain& RefChain, int M) + : Mdown(M), Nrap(Vect(RefChain.Nstrings)), Nraptot(0), + Ix2_infty(Vect(RefChain.Nstrings)), Ix2_min(Vect(RefChain.Nstrings)), Ix2_max(Vect(RefChain.Nstrings)) + { + for (int i = 0; i < RefChain.Nstrings; ++i) Nrap[i] = 0; + Nrap[0] = M; + + Nraptot = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Nraptot += Nrap[i]; + + stringstream M0out; + M0out << M; + baselabel = M0out.str(); + + // Now compute the Ix2_infty numbers + (*this).Compute_Ix2_limits(RefChain); + + // Compute dimensionality of this sub-Hilbert space: + complex ln_dimH_cx = 0.0; + for (int i = 0; i < RefChain.Nstrings; ++i) + if (Nrap[i] > 0) ln_dimH_cx += ln_Gamma(complex((Ix2_max[i] - Ix2_min[i])/2 + 2)) - ln_Gamma(complex((Ix2_max[i] - Ix2_min[i])/2 + 2 - Nrap[i])) - ln_Gamma(complex(Nrap[i] + 1)); + dimH = exp(real(ln_dimH_cx)); + } + + Heis_Base::Heis_Base (const Heis_Chain& RefChain, const Vect& Nrapidities) + : Mdown(0), Nrap(Nrapidities), Nraptot(0), + Ix2_infty(Vect(RefChain.Nstrings)), Ix2_min(Vect(RefChain.Nstrings)), Ix2_max(Vect(RefChain.Nstrings)) + { + + // Check consistency of Nrapidities vector with RefChain + + //if (RefChain.Nstrings != Nrapidities.size()) cout << "error: Nstrings = " << RefChain.Nstrings << "\tNrap.size = " << Nrapidities.size() << endl; + + if (RefChain.Nstrings != Nrapidities.size()) JSCerror("Incompatible Nrapidities vector used in Heis_Base constructor."); + + int Mcheck = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Mcheck += RefChain.Str_L[i] * Nrap[i]; + Mdown = Mcheck; + + Nraptot = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Nraptot += Nrap[i]; + + /* + // Compute id + id += Nrapidities[0]; + long long int factor = 100000LL; + for (int i = 1; i < RefChain.Nstrings; ++i) { + id += factor * Nrapidities[i]; + factor *= 100LL; + } + */ + + // Set label: + /* + stringstream baselabel_strstream; + baselabel_strstream << Nrapidities[0]; + for (int itype = 1; itype < Nrapidities.size(); ++itype) + if (Nrapidities[itype] > 0) { + baselabel_strstream << TYPESEP; + stringstream typeout; + typeout << itype; + baselabel += typeout.str(); + baselabel += 'EXCSEP'; + stringstream Mout; + Mout << Nrapidities[itype]; + baselabel += Mout.str(); + } + */ + + stringstream M0out; + M0out << Nrapidities[0]; + baselabel = M0out.str(); + for (int itype = 1; itype < Nrapidities.size(); ++itype) + if (Nrapidities[itype] > 0) { + baselabel += TYPESEP; + stringstream typeout; + typeout << itype; + baselabel += typeout.str(); + baselabel += EXCSEP; + stringstream Mout; + Mout << Nrapidities[itype]; + baselabel += Mout.str(); + } + + // Now compute the Ix2_infty numbers + (*this).Compute_Ix2_limits(RefChain); + + // Compute dimensionality of this sub-Hilbert space: + complex ln_dimH_cx = 0.0; + for (int i = 0; i < RefChain.Nstrings; ++i) + if (Nrap[i] > 0) ln_dimH_cx += ln_Gamma(complex((Ix2_max[i] - Ix2_min[i])/2 + 2)) - ln_Gamma(complex((Ix2_max[i] - Ix2_min[i])/2 + 2 - Nrap[i])) - ln_Gamma(complex(Nrap[i] + 1)); + dimH = exp(real(ln_dimH_cx)); + } + + Heis_Base::Heis_Base (const Heis_Chain& RefChain, string baselabel_ref) + : Mdown(0), Nrap(Vect(0, RefChain.Nstrings)), Nraptot(0), + Ix2_infty(Vect(RefChain.Nstrings)), Ix2_min(Vect(RefChain.Nstrings)), Ix2_max(Vect(RefChain.Nstrings)), + baselabel (baselabel_ref) + { + // Build Nrapidities vector from baselabel_ref. + // This is simply done by using the state label standard reading function after conveniently + // making baselabel into a label (as for a state with no excitations): + string label_ref = baselabel + "_0_"; + Vect > dummyOriginIx2(1); + + //cout << "Trying to build base from baselabel_ref " << baselabel_ref << "\t and label_ref " << label_ref << endl; + //State_Label_Data labeldata = Read_State_Label (label_ref, dummyOriginIx2); + State_Label_Data labeldata = Read_Base_Label (label_ref); + //cout << "Read data." << endl; + + // Initialize Nrap: + for (int i = 0; i < labeldata.type.size(); ++i) + Nrap[labeldata.type[i] ] = labeldata.M[i]; + + int Mcheck = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Mcheck += RefChain.Str_L[i] * Nrap[i]; + Mdown = Mcheck; + + Nraptot = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Nraptot += Nrap[i]; + + // Now compute the Ix2_infty numbers + (*this).Compute_Ix2_limits(RefChain); + + // Compute dimensionality of this sub-Hilbert space: + complex ln_dimH_cx = 0.0; + for (int i = 0; i < RefChain.Nstrings; ++i) + if (Nrap[i] > 0) ln_dimH_cx += ln_Gamma(complex((Ix2_max[i] - Ix2_min[i])/2 + 2)) - ln_Gamma(complex((Ix2_max[i] - Ix2_min[i])/2 + 2 - Nrap[i])) - ln_Gamma(complex(Nrap[i] + 1)); + dimH = exp(real(ln_dimH_cx)); + } + + Heis_Base& Heis_Base::operator= (const Heis_Base& RefBase) + { + if (this != & RefBase) { + Mdown = RefBase.Mdown; + Nrap = RefBase.Nrap; + Nraptot = RefBase.Nraptot; + Ix2_infty = RefBase.Ix2_infty; + Ix2_min = RefBase.Ix2_min; + Ix2_max = RefBase.Ix2_max; + dimH = RefBase.dimH; + baselabel = RefBase.baselabel; + } + return(*this); + } + + bool Heis_Base::operator== (const Heis_Base& RefBase) + { + bool answer = (Nrap == RefBase.Nrap); + + return (answer); + } + + bool Heis_Base::operator!= (const Heis_Base& RefBase) + { + bool answer = (Nrap != RefBase.Nrap); + + return (answer); + } + + void Heis_Base::Compute_Ix2_limits (const Heis_Chain& RefChain) + { + + if ((RefChain.Delta > 0.0) && (RefChain.Delta < 1.0)) { + + // Compute the Ix2_infty numbers + + DP sum1 = 0.0; + DP sum2 = 0.0; + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + sum1 = 0.0; + + for (int k = 0; k < RefChain.Nstrings; ++k) { + + sum2 = 0.0; + + sum2 += (RefChain.Str_L[j] == RefChain.Str_L[k]) ? 0.0 : 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j] * RefChain.par[k]) + - 0.5 * fabs(RefChain.Str_L[j] - RefChain.Str_L[k]) * RefChain.anis)); + sum2 += 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j] * RefChain.par[k]) + - 0.5 * (RefChain.Str_L[j] + RefChain.Str_L[k]) * RefChain.anis)); + + for (int a = 1; a < JSC::min(RefChain.Str_L[j], RefChain.Str_L[k]); ++a) + sum2 += 2.0 * 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j] * RefChain.par[k]) + - 0.5 * (fabs(RefChain.Str_L[j] - RefChain.Str_L[k]) + 2.0*a) * RefChain.anis)); + + sum1 += (Nrap[k] - ((j == k) ? 1 : 0)) * sum2; + } + + Ix2_infty[j] = (1.0/PI) * fabs(RefChain.Nsites * 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j]) + - 0.5 * RefChain.Str_L[j] * RefChain.anis)) - sum1); + + } // The Ix2_infty are now set. + + // Now compute the Ix2_max limits + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + Ix2_max[j] = int(floor(Ix2_infty[j])); // sets basic integer + + // Reject formally infinite rapidities (i.e. if Delta is root of unity) + + //cout << "Ix2_infty - Ix2_max = " << Ix2_infty[j] - Ix2_max[j] << endl; + //if (Ix2_infty[j] == Ix2_max[j]) { + //Ix2_max[j] -= 2; + //} + // If Nrap is even, Ix2_max must be odd. If odd, then even. + + if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; + + while (Ix2_max[j] > RefChain.Nsites) { + Ix2_max[j] -= 2; + } + Ix2_min[j] = -Ix2_max[j]; + } + } // if XXZ gapless + + else if (RefChain.Delta == 1.0) { + + // Compute the Ix2_infty numbers + + int sum1 = 0; + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + sum1 = 0; + + for (int k = 0; k < RefChain.Nstrings; ++k) { + + sum1 += Nrap[k] * (2 * JSC::min(RefChain.Str_L[j], RefChain.Str_L[k]) - ((j == k) ? 1 : 0)); + } + + //Ix2_infty[j] = (RefChain.Nsites - 1.0 + 2.0 * RefChain.Str_L[j] - sum1); + Ix2_infty[j] = (RefChain.Nsites + 1.0 - sum1); // to get counting right... + + } // The Ix2_infty are now set. + + // Now compute the Ix2_max limits + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + Ix2_max[j] = int(floor(Ix2_infty[j])); // sets basic integer + + // Give the correct parity to Ix2_max + + // If Nrap is even, Ix2_max must be odd. If odd, then even. + + if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; + + // If Ix2_max equals Ix2_infty, we reduce it by 2: + + if (Ix2_max[j] == int(Ix2_infty[j])) Ix2_max[j] -= 2; + + while (Ix2_max[j] > RefChain.Nsites) { + Ix2_max[j] -= 2; + } + Ix2_min[j] = -Ix2_max[j]; + } + + } // if XXX AFM + + else if (RefChain.Delta > 1.0) { + + // Compute the Ix2_infty numbers + + int sum1 = 0; + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + sum1 = 0; + + for (int k = 0; k < RefChain.Nstrings; ++k) { + + sum1 += Nrap[k] * (2 * JSC::min(RefChain.Str_L[j], RefChain.Str_L[k]) - ((j == k) ? 1 : 0)); + } + + Ix2_infty[j] = (RefChain.Nsites - 1 + 2 * RefChain.Str_L[j] - sum1); + + } // The Ix2_infty are now set. + + // Now compute the Ix2_max limits + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + Ix2_max[j] = int(floor(Ix2_infty[j])); // sets basic integer + + // Give the correct parity to Ix2_max + + // If Nrap is even, Ix2_max must be odd. If odd, then even. + + if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; + + // If Ix2_max equals Ix2_infty, we reduce it by 2: + + //if (Ix2_max[j] == Ix2_infty[j]) Ix2_max[j] -= 2; + + while (Ix2_max[j] > RefChain.Nsites) { + Ix2_max[j] -= 2; + } + + // Fudge, for strings: + //if (RefChain.Str_L[j] >= 1) Ix2_max[j] += 2; + //Ix2_max[j] += 2; + + Ix2_min[j] = -Ix2_max[j]; + } + + } // if XXZ_gpd + + } + + + + //*************************************************************************************************** + + // Function definitions: class Lambda + + Lambda::Lambda () : lambda(NULL) {} + + Lambda::Lambda (const Heis_Chain& RefChain, int M) + : Nstrings(1), Nrap(Vect(M,1)), Nraptot(M), lambda(new DP*[1]) // single type of string here + //: lambda(Vect > (1)) + { + lambda[0] = new DP[M]; + //lambda[0] = Vect (M); + + for (int j = 0; j < M; ++j) lambda[0][j] = 0.0; + } + + Lambda::Lambda (const Heis_Chain& RefChain, const Heis_Base& base) + : Nstrings(RefChain.Nstrings), Nrap(base.Nrap), Nraptot(base.Nraptot), lambda(new DP*[RefChain.Nstrings]) + //: lambda(Vect > (RefChain.Nstrings)) + { + //lambda[0] = new DP[base.Mdown]; + lambda[0] = new DP[base.Nraptot]; + for (int i = 1; i < RefChain.Nstrings; ++i) lambda[i] = lambda[i-1] + base[i-1]; + //for (int i = 0; i < RefChain.Nstrings; ++i) lambda[i] = Vect (base[i]); + + for (int i = 0; i < RefChain.Nstrings; ++i) { + for (int j = 0; j < base[i]; ++j) lambda[i][j] = 0.0; + } + } + + Lambda& Lambda::operator= (const Lambda& RefLambda) + { + if (this != &RefLambda) { + + Nstrings = RefLambda.Nstrings; + Nrap = RefLambda.Nrap; + Nraptot = RefLambda.Nraptot; + if (lambda != 0) { + delete[] (lambda[0]); + delete[] (lambda); + } + lambda = new DP*[Nstrings]; + lambda[0] = new DP[Nraptot]; + for (int i = 1; i < Nstrings; ++i) lambda[i] = lambda[i-1] + Nrap[i-1]; + for (int i = 0; i < Nstrings; ++i) + for (int j = 0; j < Nrap[i]; ++j) lambda[i][j] = RefLambda.lambda[i][j]; + + } + return(*this); + } + + Lambda::~Lambda() + { + if (lambda != 0) { + delete[] (lambda[0]); + delete[] lambda; + } + } + + + //*************************************************************************************************** + + // Function definitions: class Heis_Bethe_State + + Heis_Bethe_State::Heis_Bethe_State () + : chain(Heis_Chain()), base(Heis_Base()), //offsets(Ix2_Offsets()), + //Ix2(Ix2_Config(chain, 1)), + Ix2(Vect > (1)), + lambda(Lambda(chain, 1)), BE(Lambda(chain, 1)), diffsq(0.0), conv(0), dev(1.0), iter(0), iter_Newton(0), + E(0.0), iK(0), K(0.0), lnnorm(-100.0), //base_id(0LL), type_id(0LL), id(0LL), maxid(0LL), nparticles(0) + label("") + { + }; + + Heis_Bethe_State::Heis_Bethe_State (const Heis_Bethe_State& RefState) // copy constructor + //: chain(RefState.chain), base(RefState.base), offsets(RefState.offsets), Ix2(Ix2_Config(RefState.chain, RefState.base.Mdown)), + // lambda(Lambda(RefState.chain, RefState.base.Mdown)), BE(Lambda(RefState.chain, RefState.base.Mdown)), + : chain(RefState.chain), base(RefState.base), //offsets(RefState.offsets), + //Ix2(Ix2_Config(RefState.chain, RefState.base)), + Ix2 (RefState.Ix2), + lambda(Lambda(RefState.chain, RefState.base)), BE(Lambda(RefState.chain, RefState.base)), + diffsq(RefState.diffsq), conv(RefState.conv), dev(RefState.dev), iter(RefState.iter), iter_Newton(RefState.iter_Newton), + E(RefState.E), iK(RefState.iK), K(RefState.K), lnnorm(RefState.lnnorm), + //id(RefState.id), maxid(RefState.maxid) + //base_id(RefState.base_id), type_id(RefState.type_id), id(RefState.id), maxid(RefState.maxid), nparticles(RefState.nparticles) + label(RefState.label) + { + // copy arrays into new ones + + /* + cout << "Here in Heis constructor state" << endl; + cout << "lambda " << lambda[0][0] << endl; + cout << "lambda OK" << endl; + + cout << "RefConfig: " << endl << RefState.Ix2 << endl; + cout << "(*this).Ix2: " << endl << Ix2 << endl; + */ + for (int j = 0; j < RefState.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < RefState.base[j]; ++j) { + //Ix2[j][alpha] = RefState.Ix2[j][alpha]; // not needed anymore since Ix2 is Vect > + lambda[j][alpha] = RefState.lambda[j][alpha]; + } + } + //cout << "Heis constructor state OK" << endl; + } + + Heis_Bethe_State::Heis_Bethe_State (const Heis_Chain& RefChain, int M) + : chain(RefChain), base(RefChain, M), //offsets(base, 0LL), + //Ix2(Ix2_Config(RefChain, M)), + Ix2 (Vect > (1)), + lambda(Lambda(RefChain, M)), + BE(Lambda(RefChain, M)), diffsq(1.0), conv(0), dev(1.0), iter(0), iter_Newton(0), + E(0.0), iK(0), K(0.0), lnnorm(-100.0) + //id(0LL), maxid(0LL) + //base_id(0LL), type_id(0LL), id(0LL), maxid(offsets.maxid), nparticles(0) + { + Ix2[0] = Vect (M); + for (int j = 0; j < M; ++j) Ix2[0][j] = -(M - 1) + 2*j; + + stringstream M0out; + M0out << M; + label = M0out.str() + "_0_"; + } + + Heis_Bethe_State::Heis_Bethe_State (const Heis_Chain& RefChain, const Heis_Base& RefBase) + : chain(RefChain), base(RefBase), //offsets(RefBase, 0LL), + //Ix2(Ix2_Config(RefChain, RefBase)), + Ix2 (Vect > (RefChain.Nstrings)), + lambda(Lambda(RefChain, RefBase)), + BE(Lambda(RefChain, RefBase)), diffsq(1.0), conv(0), dev(1.0), iter(0), iter_Newton(0), E(0.0), iK(0), K(0.0), lnnorm(-100.0) + //id(0LL), maxid(0LL) + //base_id(RefBase.id), type_id(0LL), id(0LL), maxid(offsets.maxid), nparticles(0) + { + // Check that the number of rapidities is consistent with Mdown + + //cout << "Here in Heis constructor chain base" << endl; + //cout << "lambda " << lambda[0][0] << endl; + //cout << "lambda OK" << endl; + + int Mcheck = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Mcheck += base[i] * RefChain.Str_L[i]; + if (RefBase.Mdown != Mcheck) JSCerror("Wrong M from Nrapidities input vector, in Heis_Bethe_State constructor."); + + for (int i = 0; i < RefChain.Nstrings; ++i) Ix2[i] = Vect (base[i]); + + // And now for the other string types... + for (int i = 0; i < RefChain.Nstrings; ++i) { + for (int j = 0; j < base[i]; ++j) Ix2[i][j] = -(base[i] - 1) + 2*j; + } + + label = Return_State_Label (Ix2, Ix2); + } + + void Heis_Bethe_State::Set_to_Label (string label_ref, const Vect >& OriginIx2) + { + //cout << "Called Set_to_Label with label_ref " << label_ref << " on state " << (*this) << endl; + + //cout << "Check labeldata of label " << label_ref << endl; + + State_Label_Data labeldata = Read_State_Label (label_ref, OriginIx2); + + //cout << "type: " << labeldata.type << endl << "M: " << labeldata.M << endl << "nexc: " << labeldata.nexc << endl; + //for (int i = 0; i < labeldata.Ix2old.size(); ++i) cout << "Ix2old[" << i << "] = " << labeldata.Ix2old[i] << endl << "Ix2exc[" << i << "] = " << labeldata.Ix2exc[i] << endl; + + label = label_ref; + + Vect > OriginIx2ordered = OriginIx2; + for (int i = 0; i < OriginIx2.size(); ++i) OriginIx2ordered[i].QuickSort(); + + // Set all Ix2 to OriginIx2: + for (int il = 0; il < OriginIx2.size(); ++il) + for (int i = 0; i < OriginIx2[il].size(); ++i) Ix2[il][i] = OriginIx2ordered[il][i]; + + // Now set the excitations: + for (int it = 0; it < labeldata.type.size(); ++it) + for (int iexc = 0; iexc < labeldata.nexc[it]; ++iexc) + for (int i = 0; i < labeldata.M[it]; ++i) + if (Ix2[labeldata.type[it] ][i] == labeldata.Ix2old[it][iexc]) { + //cout << it << "\t" << iexc << "\t" << i << endl; + //cout << "\tSetting Ix2[" << labeldata.type[it] << "][" << i << "] to " << labeldata.Ix2exc[it][iexc] << endl; + //cout << Ix2[labeldata.type[it] ][i] << "\t" << labeldata.Ix2old[it][iexc] << "\t" << labeldata.Ix2exc[it][iexc] << endl; + Ix2[labeldata.type[it] ][i] = labeldata.Ix2exc[it][iexc]; + //cout << Ix2[labeldata.type[it] ][i] << "\t" << labeldata.Ix2old[it][iexc] << "\t" << labeldata.Ix2exc[it][iexc] << endl; + } + + //cout << "State obtained(1): " << (*this) << endl; + + // Now reorder the Ix2 to follow convention: + for (int il = 0; il < Ix2.size(); ++il) Ix2[il].QuickSort(); + + //cout << "Setting label:" << label_ref << endl << "Ix2old = " << labeldata.Ix2old[0] << endl << "Ix2exc = " << labeldata.Ix2exc[0] << endl; + //cout << "on " << OriginStateIx2ordered << endl << "giving " << Ix2 << endl; + + //cout << "State obtained(2): " << (*this) << endl; + //char a; cin >> a; + + //(*this).Set_Label_from_Ix2 (OriginStateIx2ordered); + //(*this).Set_Label_Internals_from_Ix2 (OriginStateIx2ordered); + } + + void Heis_Bethe_State::Set_Label_from_Ix2 (const Vect >& OriginIx2) + { + // This function does not assume any ordering of the Ix2. + + // ASSUMPTIONS: + // (*this) has a base already identical to base of OriginIx2 + + // First check that bases are consistent + if ((*this).chain.Nstrings != OriginIx2.size()) JSCerror("Inconsistent base sizes in Heis_Bethe_State::Set_Label_from_Ix2."); + + // Then check that the filling at each level is equal + for (int il = 0; il < (*this).chain.Nstrings; ++il) + if ((*this).base.Nrap[il] != OriginIx2[il].size()) JSCerror("Inconsistent base filling in Heis_Bethe_State::Set_Label_from_Ix2."); + + // Determine how many types of particles are present + int ntypes = 1; // level 0 always assumed present + for (int il = 1; il < chain.Nstrings; ++il) if (base.Nrap[il] > 0) ntypes++; + + // Set the state label + Vect type_ref(0, ntypes); + Vect M_ref(ntypes); + M_ref[0] = OriginIx2[0].size(); + type_ref[0] = 0; + int itype = 1; + for (int il = 1; il < chain.Nstrings; ++il) + if (base.Nrap[il] > 0) { + type_ref[itype] = il; + M_ref[itype++] = OriginIx2[il].size(); + } + Vect nexc_ref(0, ntypes); + Vect > Ix2old_ref(ntypes); + Vect > Ix2exc_ref(ntypes); + + // Count nr of particle-holes at each level: + //itype = 0; + itype = -1; + for (int il = 0; il < chain.Nstrings; ++il) { + if (il == 0 || base.Nrap[il] > 0) { + itype++; + for (int alpha = 0; alpha < base.Nrap[il]; ++alpha) + if (!OriginIx2[il].includes(Ix2[il][alpha])) nexc_ref[itype] += 1; + Ix2old_ref[itype] = Vect(JSC::max(nexc_ref[itype], 1)); + Ix2exc_ref[itype] = Vect(JSC::max(nexc_ref[itype], 1)); + int nexccheck = 0; + for (int alpha = 0; alpha < base.Nrap[il]; ++alpha) + if (!OriginIx2[il].includes(Ix2[il][alpha])) Ix2exc_ref[itype][nexccheck++] = Ix2[il][alpha]; + if (nexccheck != nexc_ref[itype]) { + cout << "il = " << il << "\titype = " << itype << "\tnexccheck = " << nexccheck << "\tnexc_ref[itype] = " << nexc_ref[itype] << endl; + cout << OriginIx2[il] << endl << Ix2[il] << endl; + JSCerror("Counting excitations wrong (1) in Heis_Bethe_State::Set_Label_from_Ix2."); + } + nexccheck = 0; + for (int alpha = 0; alpha < base.Nrap[il]; ++alpha) + if (!Ix2[il].includes (OriginIx2[il][alpha])) Ix2old_ref[itype][nexccheck++] = OriginIx2[il][alpha]; + if (nexccheck != nexc_ref[itype]) { + JSCerror("Counting excitations wrong (2) in Heis_Bethe_State::Set_Label_from_Ix2."); + } + // Now order the Ix2old_ref and Ix2exc_ref: + Ix2old_ref[itype].QuickSort(); + Ix2exc_ref[itype].QuickSort(); + } // if (il == 0 || ... + } // for il + + State_Label_Data labeldata(type_ref, M_ref, nexc_ref, Ix2old_ref, Ix2exc_ref); + + label = Return_State_Label (labeldata, OriginIx2); + } + + bool Heis_Bethe_State::Check_Symmetry () + { + // Checks whether the I's are symmetrically distributed. + + bool symmetric_state = true; + int arg, test1, test3; + + //if (chain.Delta <= 1.0) { + + for (int j = 0; j < chain.Nstrings; ++j) { + test1 = 0; + test3 = 0; + for (int alpha = 0; alpha < base[j]; ++alpha) { + arg = (Ix2[j][alpha] != chain.Nsites) ? Ix2[j][alpha] : 0; // since Ix2 = N is same as Ix2 = -N by periodicity, this is symmetric. + test1 += arg; + test3 += arg * arg * arg; // to make sure that all I's are symmetrical... + } + if (!(symmetric_state && (test1 == 0) && (test3 == 0))) symmetric_state = false; + } + //} + /* + else if (chain.Delta > 1.0) { + // For the gapped antiferromagnet, we check symmetry *excluding* the Ix2_max values + // The state is then inadmissible is symmetric && -Ix2_max occupied + for (int j = 0; j < chain.Nstrings; ++j) { + test1 = 0; + test3 = 0; + for (int alpha = 0; alpha < base[j]; ++alpha) { + arg = (Ix2[j][alpha] != chain.Nsites + && abs(Ix2[j][alpha]) != base.Ix2_max[j]) ? Ix2[j][alpha] : 0; // since Ix2 = N is same as Ix2 = -N by periodicity, this is symmetric. + test1 += arg; + test3 += arg * arg * arg; // to make sure that all I's are symmetrical... + } + if (!(symmetric_state && (test1 == 0) && (test3 == 0))) symmetric_state = false; + } + } + */ + + return symmetric_state; + } + + + // SOLVING BETHE EQUATIONS: + + void Heis_Bethe_State::Compute_diffsq () + { + DP maxterm = 0.0; + int jmax, alphamax; + + diffsq = 0.0; + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + diffsq += pow(BE[j][alpha], 2.0); + if (pow(BE[j][alpha], 2.0)/chain.Nsites > maxterm) { + jmax = j; + alphamax = alpha; + maxterm = pow(BE[j][alpha], 2.0)/chain.Nsites; + } + } + diffsq /= DP(chain.Nsites); + } + + void Heis_Bethe_State::Find_Rapidities (bool reset_rapidities) + { + // This function finds the rapidities of the eigenstate + + //cout << endl << "Find_Rapidities called: " << (*this) << endl; + + lnnorm = -100.0; // sentinel value, recalculated if Newton method used in the last step of iteration. + + Lambda lambda_ref(chain, base); + + diffsq = 1.0; + + if (reset_rapidities) + (*this).Set_Free_lambdas(); + //cout << endl << "After Set_Free_lambdas: " << (*this) << endl; + (*this).Compute_BE(); + //cout << endl << "After Compute_BE: " << (*this) << endl; + + + iter = 0; + iter_Newton = 0; + + // Start with conventional iterations + + DP iter_prec = 1.0e-2/DP(chain.Nsites); + DP iter_factor = 0.99; + DP iter_factor_extrap = 0.99; + + clock_t extrap_start; + clock_t extrap_stop; + clock_t Newton_start; + clock_t Newton_stop; + int iter_extrap_start, iter_extrap_stop, iter_Newton_start, iter_Newton_stop; + DP diffsq_start; + DP diffsq_extrap_start, diffsq_extrap_stop, diffsq_Newton_start, diffsq_Newton_stop; + + bool straight_improving = true; + bool extrap_improving = true; + bool Newton_improving = true; + + DP diffsq_iter_aim = 1.0e-5; + + bool info_findrap = false; + + if (info_findrap) cout << "Find_Rapidities called for state with label " << (*this).label << endl; + + while (diffsq > chain.prec && (iter < 500 && iter_Newton < 100 + || straight_improving || extrap_improving || Newton_improving)) { + + // If we haven't reset, first try a few Newton steps... + //if (!reset_rapidities && iter_Newton == 0) (*this).Solve_BAE_Newton (chain.prec, 10); + + if (!Newton_improving) diffsq_iter_aim *= 1.0e-5; + + // Start with a few straight iterations: + + if (!straight_improving) iter_factor *= 2.0/3.0; + else iter_factor = 0.99; + + diffsq_start = diffsq; + do { + extrap_start = clock(); + iter_extrap_start = iter; + diffsq_extrap_start = diffsq; + + if (diffsq > iter_prec) (*this).Solve_BAE_straight_iter (iter_prec, 10, iter_factor); + + extrap_stop = clock(); + iter_extrap_stop = iter; + diffsq_extrap_stop = diffsq; + + iter_prec = diffsq * 0.1; + + if (info_findrap) cout << "Straight iter: iter " << iter << "\titer_factor " << iter_factor << "\tdiffsq " << diffsq << endl; + + //} while (diffsq > chain.prec && !is_nan(diffsq) && diffsq_extrap_stop/diffsq_extrap_start < 0.01); + } while (diffsq > diffsq_iter_aim && !is_nan(diffsq) && diffsq_extrap_stop/diffsq_extrap_start < 0.01); + + straight_improving = (diffsq < diffsq_start); + + // Now try to extrapolate to infinite nr of iterations: + + if (!extrap_improving) iter_factor_extrap *= 2.0/3.0; + else iter_factor_extrap = 0.99; + + diffsq_start = diffsq; + + if (diffsq > chain.prec) + do { + extrap_start = clock(); + iter_extrap_start = iter; + diffsq_extrap_start = diffsq; + + if (diffsq > iter_prec) (*this).Solve_BAE_extrap (iter_prec, 10, iter_factor_extrap); + + extrap_stop = clock(); + iter_extrap_stop = iter; + diffsq_extrap_stop = diffsq; + + iter_prec = diffsq * 0.1; + + if (info_findrap) cout << "Extrap: iter " << iter << "\tdiffsq " << diffsq << endl; + + } while (diffsq > chain.prec && !is_nan(diffsq) && diffsq_extrap_stop/diffsq_extrap_start < 0.01); + + extrap_improving = (diffsq < diffsq_extrap_start); + + // Now try Newton + + diffsq_Newton_start = diffsq; + if (diffsq > chain.prec) + do { + Newton_start = clock(); + iter_Newton_start = iter_Newton; + diffsq_Newton_start = diffsq; + + if (diffsq > iter_prec) (*this).Solve_BAE_Newton (chain.prec, 5); + + Newton_stop = clock(); + iter_Newton_stop = iter_Newton; + diffsq_Newton_stop = diffsq; + + if (info_findrap) cout << "Newton: iter_Newton " << iter_Newton << "\tdiffsq " << diffsq << endl; + + } while (diffsq > chain.prec && !is_nan(diffsq) && diffsq_Newton_stop/diffsq_Newton_start < 0.01); + + Newton_improving = (diffsq < diffsq_Newton_start); + + //if (diffsq > iter_prec) (*this).Solve_BAE_smackdown (0.1 * diffsq, 1); + + //cout << "Before silk: diffsq = " << diffsq << endl; + + // If none of the methods are improving the result, try the silk gloves... + if (!(straight_improving || extrap_improving || Newton_improving)) { + if (info_findrap) cout << "Before silk gloves: diffsq " << diffsq << endl; + (*this).Solve_BAE_with_silk_gloves (chain.prec, chain.Nsites, 0.9); + if (info_findrap) cout << "After silk gloves: diffsq " << diffsq << endl; + } + + if (is_nan(diffsq)) { + (*this).Set_Free_lambdas(); // good start if we've messed up with Newton + diffsq = 1.0; + } + + iter_prec *= 1.0e-4; + iter_prec = JSC::max(iter_prec, chain.prec); + + } // while (diffsq > chain.prec && (iter < 300 && iter_Newton < 50... + + // Check convergence: + + //cout << "Check_Rapidities: " << (*this).Check_Rapidities() << endl; + + //conv = (diffsq < chain.prec && (*this).Check_Rapidities() && ((*this).String_delta() < HEIS_deltaprec)) ? 1 : 0; + conv = (diffsq < chain.prec && (*this).Check_Rapidities()) ? 1 : 0; + dev = (*this).String_delta(); + + //cout << "String delta: " << (*this).String_delta() << "\tBoolean: " << ((*this).String_delta() < HEIS_deltaprec) << endl; + + return; + } + + void Heis_Bethe_State::Solve_BAE_bisect (int j, int alpha, DP req_prec, int itermax) + { + // Finds the root lambda[j][alpha] to precision req_prec using bisection + + DP prec_obtained = 1.0; + int niter_here = 0; + + DP lambdajalphastart = lambda[j][alpha]; + DP lambdamax = (fabs(chain.Delta) <= 1.0 ? pow(log(chain.Nsites), 1.1) : PI - 1.0e-4); + + // Find values of lambda such that BE changes sign: + DP lambdaleft = -lambdamax; + DP lambdamid = 0.0; + DP lambdaright = lambdamax; + + DP BEleft, BEmid, BEright; + lambda[j][alpha] = lambdaleft; + (*this).Compute_BE (j, alpha); + //cout << "lambda: " << lambda[j][alpha] << "\ttanhlambda: " << tanh(lambda[j][alpha]) << endl; + BEleft= BE[j][alpha]; + lambda[j][alpha] = lambdaright; + (*this).Compute_BE (j, alpha); + BEright= BE[j][alpha]; + + if (BEleft * BEright > 0.0) { + lambda[j][alpha] = lambdajalphastart; + //cout << lambdaleft << "\t" << lambdaright << "\t" << BEleft << "\t" << BEright << endl; + //cout << "Could not bisect BE[" << j << "][" << alpha << "]" << endl; + return; + } + + while (prec_obtained > req_prec && niter_here < itermax) { + + lambdamid = 0.5 * (lambdaleft + lambdaright); + lambda[j][alpha] = lambdamid; + (*this).Compute_BE (j, alpha); + BEmid = BE[j][alpha]; + + //cout << "niter_here = " << niter_here << "\t" << lambdaleft << "\t" << lambdamid << "\t" << lambdaright << endl; + //cout << BEleft << "\t" << BEmid << "\t" << BEright << endl; + + if (BEmid * BEleft < 0.0) { // root is to the left of mid + lambdaright = lambdamid; + BEright = BEmid; + } + + else if (BEmid * BEright < 0.0) { // root is to the right of mid + lambdaleft = lambdamid; + BEleft = BEmid; + } + + else if (BEmid * BEmid < req_prec) return; + + else { + //cout << lambdaleft << "\t" << lambdamid << "\t" << lambdaright << endl; + //cout << BEleft << "\t" << BEmid << "\t" << BEright << endl; + //JSCerror("Problem in Solve_BAE_bisect."); + return; // this procedure has failed + } + + prec_obtained = BEmid * BEmid; + niter_here++; + + //cout << "bisect: " << lambdaleft << "\t" << lambdaright << "\t" << BEleft << "\t" << BEright << "\t" << prec_obtained << "\t" << req_prec << endl; + } + + return; + } + + void Heis_Bethe_State::Iterate_BAE (DP iter_factor) + { + //DP lambda_old; + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) + { + //lambda_old = lambda[j][alpha]; + //lambda[j][alpha] = Iterate_BAE (j, alpha); + lambda[j][alpha] += iter_factor * (Iterate_BAE (j, alpha) - lambda[j][alpha]); + //cout << j << "\t" << alpha << "\t" << Ix2[j][alpha] << "\t" << lambda_old << "\t" << lambda[j][alpha] << "\t"; + //if (j > 0) cout << j << "\t" << alpha << "\t" << Ix2[j][alpha] << "\t" << lambda_old << "\t" << lambda[j][alpha] << endl; + } + } + + iter++; + + (*this).Compute_BE(); + (*this).Compute_diffsq(); + } + + void Heis_Bethe_State::Solve_BAE_straight_iter (DP straight_prec, int max_iter, DP iter_factor) + { + // This function attempts to get convergence diffsq <= straight_prec in at most max_iter steps. + + // If this fails, the lambda's are reset to original values, defined as... + + Lambda lambda_ref(chain, base); + DP diffsq_ref = 1.0; + + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda_ref[j][alpha] = lambda[j][alpha]; + diffsq_ref = diffsq; + + // Now begin solving... + + diffsq = 1.0; + int iter_done_here = 0; + + while ((diffsq > straight_prec) && (max_iter > iter_done_here)) { + + //cout << "BEFORE ITERATION" << endl << (*this) << endl << endl; + (*this).Iterate_BAE(iter_factor); + //cout << "ITERATION " << iter_done_here << endl << (*this) << endl << endl; + iter_done_here++; + } + + if ((diffsq > diffsq_ref) || (is_nan(diffsq))) { + // This procedure has failed. We reset everything to begin values. + //cout << "Straight iter failed: resetting." << endl; + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda[j][alpha] = lambda_ref[j][alpha]; + (*this).Compute_BE(); + diffsq = diffsq_ref; + } + } + + void Heis_Bethe_State::Solve_BAE_extrap (DP extrap_prec, int max_iter_extrap, DP iter_factor) + { + // This function attempts to get convergence diffsq <= extrap_prec in at most max_iter_extrap steps. + + // If this fails, the lambda's are reset to original values, defined as... + + Lambda lambda_ref(chain, base); + DP diffsq_ref = 1.0; + + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda_ref[j][alpha] = lambda[j][alpha]; + diffsq_ref = diffsq; + + // Now begin solving... + + diffsq = 1.0; + DP diffsq1, diffsq2, diffsq3, diffsq4; + + int iter_done_here = 0; + + Lambda lambda1(chain, base); + Lambda lambda2(chain, base); + Lambda lambda3(chain, base); + Lambda lambda4(chain, base); + Lambda lambda5(chain, base); + + while ((diffsq > extrap_prec) && (max_iter_extrap > iter_done_here)) { + + (*this).Iterate_BAE(iter_factor); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda1[j][alpha] = lambda[j][alpha]; + diffsq1 = diffsq; + (*this).Iterate_BAE(iter_factor); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda2[j][alpha] = lambda[j][alpha]; + diffsq2 = diffsq; + (*this).Iterate_BAE(iter_factor); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda3[j][alpha] = lambda[j][alpha]; + diffsq3 = diffsq; + (*this).Iterate_BAE(iter_factor); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda4[j][alpha] = lambda[j][alpha]; + diffsq4 = diffsq; + //(*this).Iterate_BAE(iter_factor); + //for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda5[j][alpha] = lambda[j][alpha]; + //diffsq5 = diffsq; + + iter_done_here += 4; + + // now extrapolate the result to infinite number of iterations... + + if (diffsq4 < diffsq3 && diffsq3 < diffsq2 && diffsq2 < diffsq1 && diffsq1 < diffsq_ref) { + + Vect_DP rap(0.0, 4); + Vect_DP oneoverP(0.0, 4); + DP deltalambda = 0.0; + + for (int i = 0; i < 4; ++i) oneoverP[i] = 1.0/(1.0 + i*i); + + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) { + rap[0] = lambda1[j][alpha]; + rap[1] = lambda2[j][alpha]; + rap[2] = lambda3[j][alpha]; + rap[3] = lambda4[j][alpha]; + //rap[4] = lambda5[j][alpha]; + + polint (oneoverP, rap, 0.0, lambda[j][alpha], deltalambda); + + //cout << j << "\t" << alpha << "\t" << rap << "\t" << lambda[j][alpha] << "\t" << deltalambda << endl; + } + + // Iterate once to stabilize result + (*this).Iterate_BAE(iter_factor); + } + + } // ((diffsq > extrap_prec) && (max_iter_extrap > iter_done_here)) + + if ((diffsq >= diffsq_ref) || (is_nan(diffsq))) { + // This procedure has failed. We reset everything to begin values. + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda[j][alpha] = lambda_ref[j][alpha]; + (*this).Compute_BE(); + diffsq = diffsq_ref; + } + + return; + } + + void Heis_Bethe_State::Solve_BAE_with_silk_gloves (DP silk_prec, int max_iter_silk, DP iter_factor) + { + // Logic: do iterations, but change only one rapidity at a time. + // This is slow, and called only if straight_iter, extrap and Newton methods don't work. + + // If this fails, the lambda's are reset to original values, defined as... + + Lambda lambda_ref(chain, base); + DP diffsq_ref = 1.0; + + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda_ref[j][alpha] = lambda[j][alpha]; + diffsq_ref = diffsq; + + // Now begin solving... + + diffsq = 1.0; + int iter_done_here = 0; + + while ((diffsq > silk_prec) && (max_iter_silk > iter_done_here)) { + + // Find the highest `deviant' rapidity + int jmax = 0; + int alphamax = 0; + DP absBEmax = 0.0; + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + //cout << j << "\t" << alpha << "\t" << BE[j][alpha] << endl; + if (fabs(BE[j][alpha]) > absBEmax) { + jmax = j; + alphamax = alpha; + absBEmax = fabs(BE[j][alpha]); + } + } + + //cout << "jmax = " << jmax << "\talphamax = " << alphamax << "\t" << lambda[jmax][alphamax] << "\tBE before " << BE[jmax][alphamax] << endl; + // Now recalculate this max deviant rapidity, + //cout << lambda[jmax][alphamax] << "\t" << Iterate_BAE (jmax, alphamax) << endl; + + /* + DP dlambda = 0.0; + DP prevBEmax = 0.0; + do { + dlambda = Iterate_BAE (jmax, alphamax) - lambda[jmax][alphamax]; + lambda[jmax][alphamax] += iter_factor * dlambda; + prevBEmax = BE[jmax][alphamax]; + (*this).Compute_BE(); + iter_done_here++; + cout << "jmax = " << jmax << "\talphamax = " << alphamax << "\t" << lambda[jmax][alphamax] << "\t" << dlambda << "\tBE during " << BE[jmax][alphamax] << endl; + } while (dlambda * dlambda > silk_prec && fabs(BE[jmax][alphamax]) < fabs(prevBEmax) && max_iter_silk > iter_done_here); + + */ + + Solve_BAE_bisect (jmax, alphamax, silk_prec, max_iter_silk); + iter_done_here++; + + //cout << "jmax = " << jmax << "\talphamax = " << alphamax << "\t" << lambda[jmax][alphamax] << "\tBE after " << BE[jmax][alphamax] << endl; + + // and reset all important arrays. + (*this).Compute_diffsq(); + } + + //cout << "Silk gloves: diffsq from " << diffsq_ref << "\tto " << diffsq << endl; + + if ((diffsq > diffsq_ref) || (is_nan(diffsq))) { + // This procedure has failed. We reset everything to begin values. + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda[j][alpha] = lambda_ref[j][alpha]; + (*this).Compute_BE(); + diffsq = diffsq_ref; + } + + return; + } + + /* + void Heis_Bethe_State::BAE_smackdown (DP max_allowed) + { + // Re-solves for all rapidities lambda[j][alpha] such that BE[j][alpha]^2/N > max_allowed. + // Assumes that BE[][] is up-to-date. + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) + if (pow(BE[j][alpha], 2.0)/chain.Nsites > max_allowed) (*this).Solve_BAE (j, alpha, max_allowed, 100); + } + + void Heis_Bethe_State::Solve_BAE_smackdown (DP max_allowed, int maxruns) + { + int runs_done = 0; + + (*this).Compute_BE(); + (*this).Compute_diffsq(); + + while (diffsq > chain.prec && diffsq > max_allowed && runs_done < maxruns) { + (*this).BAE_smackdown (max_allowed); + (*this).Compute_BE(); + (*this).Compute_diffsq(); + runs_done++; + } + } + */ + + void Heis_Bethe_State::Iterate_BAE_Newton () + { + // does one step of a Newton method on the rapidities... + // Assumes that BE[j][alpha] have been computed + + Vect_CX dlambda (0.0, base.Nraptot); // contains delta lambda computed from Newton's method + SQMat_CX Gaudin (0.0, base.Nraptot); + Vect_INT indx (base.Nraptot); + + Lambda lambda_ref(chain, base); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda_ref[j][alpha] = lambda[j][alpha]; + + (*this).Build_Reduced_Gaudin_Matrix (Gaudin); + + int index = 0; + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + dlambda[index] = - BE[j][alpha] * chain.Nsites; + index++; + } + + DP d; + + try { + ludcmp_CX (Gaudin, indx, d); + lubksb_CX (Gaudin, indx, dlambda); + } + + catch (Divide_by_zero) { + diffsq = log(-1.0); // reset to nan to stop Newton method + return; + } + + //cout << iter_Newton << "\t" << dlambda << endl; + + // Regularize dlambda: max step is +-1.0 to prevent rapidity flying off into outer space. + for (int i = 0; i < base.Nraptot; ++i) if (fabs(real(dlambda[i])) > 1.0) dlambda[i] = 0.0;//(real(dlambda[i]) > 0) ? 1.0 : -1.0; + + index = 0; + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + lambda[j][alpha] = lambda_ref[j][alpha] + real(dlambda[index]); + //cout << j << "\t" << alpha << "\t" << dlambda[index] << "\t" << lambda_ref[j][alpha] << "\t" << lambda[j][alpha] << endl; + index++; + } + + (*this).Compute_BE(); + (*this).Iterate_BAE(1.0); + (*this).Compute_diffsq(); + + // if we've converged, calculate the norm here, since the work has been done... + + if (diffsq < chain.prec) { + lnnorm = 0.0; + for (int j = 0; j < base.Nraptot; j++) lnnorm += log(abs(Gaudin[j][j])); + } + + iter_Newton++; + + return; + } + + void Heis_Bethe_State::Solve_BAE_Newton (DP Newton_prec, int max_iter_Newton) + { + // This function attempts to get convergence diffsq <= Newton_prec in at most max_iter_Newton steps. + + // The results are accepted if diffsq has decreased, otherwise the lambda's are reset to original values, defined as... + + Lambda lambda_ref(chain, base); + DP diffsq_ref = 1.0; + + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda_ref[j][alpha] = lambda[j][alpha]; + diffsq_ref = diffsq; + + // Now begin solving... + + int iter_done_here = 0; + + while ((diffsq > Newton_prec) && (diffsq < 10.0) && (!is_nan(diffsq)) && (iter_done_here < max_iter_Newton)) { + (*this).Iterate_BAE_Newton(); + iter_done_here++; + (*this).Iterate_BAE(1.0); + } + + if ((diffsq > diffsq_ref) || (is_nan(diffsq))) { + // This procedure has failed. We reset everything to begin values. + //cout << "Newton: failed, resetting." << "\t" << diffsq << endl; + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda[j][alpha] = lambda_ref[j][alpha]; + (*this).Compute_BE(); + diffsq = diffsq_ref; + } + + return; + } + + void Heis_Bethe_State::Compute_lnnorm () + { + if (true || lnnorm == -100.0) { // else norm already calculated by Newton method + // Actually, compute anyway to increase accuracy ! + + SQMat_CX Gaudin_Red(base.Nraptot); + + (*this).Build_Reduced_Gaudin_Matrix(Gaudin_Red); + /* + cout << endl << "Gaudin matrix: " << endl; + + for (int j = 0; j < Gaudin_Red.size(); ++j) { + for (int k = 0; k < Gaudin_Red.size(); ++k) cout << Gaudin_Red[j][k] << " "; + cout << endl; + } + cout << endl << endl; + */ + complex lnnorm_check = lndet_LU_CX_dstry(Gaudin_Red); + //cout << "Calculated lnnorm = " << lnnorm_check; + + //lnnorm = real(lndet_LU_CX_dstry(Gaudin_Red)); + lnnorm = real(lnnorm_check); + } + + return; + } + + void Heis_Bethe_State::Compute_Momentum () + { + int sum_Ix2 = 0; + DP sum_M = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + sum_M += 0.5 * (1.0 + chain.par[j]) * base[j]; + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum_Ix2 += Ix2[j][alpha]; + } + } + + iK = (chain.Nsites/2) * int(sum_M + 0.1) - (sum_Ix2/2); // + 0.1: for safety... + + while (iK >= chain.Nsites) iK -= chain.Nsites; + while (iK < 0) iK += chain.Nsites; + + K = PI * sum_M - PI * sum_Ix2/chain.Nsites; + + while (K >= 2.0*PI) K -= 2.0*PI; + while (K < 0.0) K += 2.0*PI; + + return; + } + + void Heis_Bethe_State::Compute_All (bool reset_rapidities) // solves BAE, computes E, K and lnnorm + { + (*this).Find_Rapidities (reset_rapidities); + if (conv == 1) { + (*this).Compute_Energy (); + (*this).Compute_Momentum (); + (*this).Compute_lnnorm (); + } + return; + } + + /* + bool Heis_Bethe_State::Boost_Momentum (int iKboost) + { + if (iKboost == 0) return(true); // done + + Ix2_Offsets offsets_here = offsets; + + bool success = false; + + if (iKboost < 0) + success = offsets_here.Add_Boxes_From_Lowest(-iKboost, 0); // add boxes in even sectors to decrease iK + + else if (iKboost > 0) + success = offsets_here.Add_Boxes_From_Lowest(iKboost, 1); // add boxes in odd sectors to increase iK + + if (success) (*this).Set_Ix2_Offsets(offsets_here); + + return(success); + } + */ + + + void Heis_Bethe_State::Set_to_Closest_Matching_Ix2_fixed_Base (const Heis_Bethe_State& StateToMatch) + { + // Given a state with given Ix2 distribution, set the Ix2 to closest match. + // The base of (*this) is fixed, and does not necessarily match that of StateToMatch. + + //cout << "Matching Ix2 for base " << (*this).base.baselabel << " from base " << StateToMatch.base.baselabel << endl; + + if ((*this).chain != StateToMatch.chain) + JSCerror("Heis_Bethe_State::Find_Closest_Matching_Ix2_fixed_Base: trying to match Ix2 for two states with different chains."); + + // Check level by level, match quantum numbers from center up. + for (int il = 0; il < chain.Nstrings; ++il) { + // Careful: if parity of Nraps is different, the Ix2 live on different lattices! + int Ix2shift = ((StateToMatch.base.Nrap[il] - (*this).base.Nrap[il]) % 2); + if (StateToMatch.base.Nrap[il] >= (*this).base.Nrap[il]) { + // We match the rapidities from the center towards the sides. + int ashift = (StateToMatch.base.Nrap[il] - (*this).base.Nrap[il])/2; + for (int a = 0; a < (*this).base.Nrap[il]; ++a) + (*this).Ix2[il][a] = StateToMatch.Ix2[il][a + ashift] + Ix2shift; + } + else { // There are less Ix2 in StateToMatch than in (*this) + // We start by filling all the (*this) Ix2 symmetrically from the middle. + // We thereafter start from the left and identify half of the StateToMatch Ix2 with the leftmost (*this)Ix2 + // and then do the same thing from the right. + for (int a = 0; a < (*this).base.Nrap[il]; ++a) + (*this).Ix2[il][a] = -(*this).base.Nrap[il] + 1 + 2*a; + int nleft = StateToMatch.base.Nrap[il]/2; + for (int a = 0; a < nleft; ++a) + if (StateToMatch.Ix2[il][a] - Ix2shift < (*this).Ix2[il][a]) (*this).Ix2[il][a] = StateToMatch.Ix2[il][a] - Ix2shift; + for (int a = 0; a < StateToMatch.base.Nrap[il] - 1 - nleft; ++a) + if (StateToMatch.Ix2[il][StateToMatch.base.Nrap[il] - 1 - a] - Ix2shift > (*this).Ix2[il][(*this).base.Nrap[il] - 1 - a]) + (*this).Ix2[il][(*this).base.Nrap[il] - 1 - a] = StateToMatch.Ix2[il][StateToMatch.base.Nrap[il] - 1 - a] - Ix2shift; + } + } // for il + + //cout << "StateToMatch:" << endl << StateToMatch << endl << "MatchingState:" << endl << (*this) << endl; + + } + + + std::ostream& operator<< (std::ostream& s, const Heis_Bethe_State& state) + { + // sends all the state data to output stream + + s << endl << "******** Chain with Delta = " << state.chain.Delta << " Nsites = " << state.chain.Nsites << " Mdown = " << state.base.Mdown + //<< ": eigenstate with base_id " << state.base_id << ", type_id " << state.type_id << " id " << state.id << " maxid " << state.maxid << endl + << ": eigenstate with label " << state.label << endl + << "E = " << state.E << " K = " << state.K << " iK = " << state.iK << " lnnorm = " << state.lnnorm << endl + << "conv = " << state.conv << " dev = " << state.dev << " iter = " << state.iter << " iter_Newton = " << state.iter_Newton << "\tdiffsq " << state.diffsq << endl; + + for (int j = 0; j < state.chain.Nstrings; ++j) { + if (state.base.Nrap[j] > 0) { + s << "Type " << j << " Str_L = " << state.chain.Str_L[j] << " par = " << state.chain.par[j] << " M_j = " << state.base.Nrap[j] + << " Ix2_infty = " << state.base.Ix2_infty[j] << " Ix2_min = " << state.base.Ix2_min[j] << " Ix2_max = " << state.base.Ix2_max[j] << endl; + Vect_INT qnumbers(state.base.Nrap[j]); + Vect_DP rapidities(state.base.Nrap[j]); + for (int alpha = 0; alpha < state.base.Nrap[j]; ++alpha) { + qnumbers[alpha] = state.Ix2[j][alpha]; + rapidities[alpha] = state.lambda[j][alpha]; + } + qnumbers.QuickSort(); + rapidities.QuickSort(); + + s << "Ix2 quantum numbers: " << endl; + for (int alpha = 0; alpha < state.base.Nrap[j]; ++alpha) s << qnumbers[alpha] << " "; + s << endl; + s << "Rapidities: " << endl; + for (int alpha = 0; alpha < state.base.Nrap[j]; ++alpha) s << rapidities[alpha] << " "; + s << endl; + } + } + s << endl; + + return s; + + } + + +} // namespace JSC diff --git a/src/HEIS/Heis_Chem_Pot.cc b/src/HEIS/Heis_Chem_Pot.cc new file mode 100644 index 0000000..0d9764f --- /dev/null +++ b/src/HEIS/Heis_Chem_Pot.cc @@ -0,0 +1,142 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/HEIS/Heis_Chem_Pot.cc + +Purpose: calculates the chemical potential. + + +***********************************************************/ + +#include "JSC.h" + +namespace JSC { + + DP Ezero (DP Delta, int N, int M) + { + // Returns the energy of the ground state with M down spins + + if (M < 0 || M > N/2) JSCerror("M out of bounds in Ezero."); + + DP E = -1.0; // sentinel value + + if (M == 0) E = N * Delta/4.0; + + else { + + Heis_Chain BD1(1.0, Delta, 0.0, N); + + Vect_INT Nrapidities_groundstate(0, BD1.Nstrings); + + Nrapidities_groundstate[0] = M; + + Heis_Base baseconfig_groundstate(BD1, Nrapidities_groundstate); + + if ((Delta > 0.0) && (Delta < 1.0)) { + XXZ_Bethe_State groundstate(BD1, baseconfig_groundstate); + groundstate.Compute_All(true); + E = groundstate.E; + } + + else if (Delta == 1.0) { + XXX_Bethe_State groundstate(BD1, baseconfig_groundstate); + groundstate.Compute_All(true); + E = groundstate.E; + } + + else if (Delta > 1.0) { + XXZ_gpd_Bethe_State groundstate(BD1, baseconfig_groundstate); + groundstate.Compute_All(true); + E = groundstate.E; + } + + else JSCerror("Anisotropy out of bounds in Ezero."); + } + + return(E); + } + + DP H_vs_M (DP Delta, int N, int M) + { + // Assumes dE/dM = 0 = dE_0/dM + h, with dE_0/dM = E_0(M) - E_0 (M - 1) + + DP H = 0.0; + + if (2*M == N) H = 0.0; + + else if (Delta <= 1.0) H = Ezero (Delta, N, M - 1) - Ezero (Delta, N, M); + + return(H); + } + + DP HZmin (DP Delta, int N, int M, Vect_DP& Ezero_ref) + { + if (M < 0 || M > N/2 - 1) { + cout << "M = " << M << endl; + JSCerror("M out of bounds in HZmin."); + } + + if (Ezero_ref[M] == -1.0) Ezero_ref[M] = Ezero(Delta, N, M); + if (Ezero_ref[M + 1] == -1.0) Ezero_ref[M + 1] = Ezero(Delta, N, M + 1); + + return(Ezero_ref[M] - Ezero_ref[M + 1]); + } + + int M_vs_H (DP Delta, int N, DP HZ) + { + // Returns the value of M for given field HZ + + if (HZ < 0.0) JSCerror("Please use a positive field in M_vs_H."); + + else if (HZ == 0.0) return(N/2); + + // Here, -1.0 is a sentinel value. + Vect_DP Ezero(-1.0, N/2 + 1); // contains the GSE[M]. + + // We look for M s.t. HZmin[M] < HZ <= HZmin[M + 1] + + int M_actual = N/4; // start somewhere in middle + int M_step = N/8 - 1; // step + DP HZmin_actual = 0.0; + DP HZmax_actual = 0.0; + bool M_found = false; + + if (HZ >= 1.0 + Delta) M_actual = 0; // saturation + + else { + + HZmin_actual = HZmin (Delta, N, M_actual, Ezero); + HZmax_actual = HZmin (Delta, N, M_actual - 1, Ezero); + + while (!M_found) { + + if (HZmin_actual > HZ) M_actual += M_step; + else if (HZmax_actual <= HZ) M_actual -= M_step; + + M_step = (M_step + 1)/2; + + HZmin_actual = HZmin (Delta, N, M_actual, Ezero); + HZmax_actual = HZmin (Delta, N, M_actual - 1, Ezero); + + M_found = (HZmin_actual < HZ && HZ <= HZmax_actual); + + //cout << "M_actual = " << M_actual << "\tM_step = " << M_step + // << "\tHZmin_actual = " << HZmin_actual << "\tHZmax_actual = " << HZmax_actual << "\tHZ = " << HZ << "\t" << M_found << endl; + } + } + //cout << "M found = " << M_actual << "\tHZmax = " << Ezero[M_actual] - Ezero[M_actual + 1] << "\tHZmin = " << Ezero[M_actual - 1] - Ezero[M_actual] << endl; + + return(M_actual); + } + + DP Chemical_Potential (const Heis_Bethe_State& AveragingState) + { + return(-H_vs_M (AveragingState.chain.Delta, AveragingState.chain.Nsites, AveragingState.base.Mdown)); // - sign since E_{M+1} - E_M = -H + } + +} diff --git a/src/HEIS/Heis_Matrix_Element_Contrib.cc b/src/HEIS/Heis_Matrix_Element_Contrib.cc new file mode 100644 index 0000000..bf7821e --- /dev/null +++ b/src/HEIS/Heis_Matrix_Element_Contrib.cc @@ -0,0 +1,376 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/HEIS/Heis_Matrix_Element_Contrib.cc + +Purpose: handles the generic call for a matrix element. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + //DP Compute_Matrix_Element_Contrib (char whichDSF, bool fixed_iK, XXZ_Bethe_State& LeftState, + //XXZ_Bethe_State& RightState, DP Chem_Pot, fstream& DAT_outfile) + //DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXZ_Bethe_State& LeftState, + // XXZ_Bethe_State& RightState, DP Chem_Pot, fstream& DAT_outfile) + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXZ_Bethe_State& LeftState, + XXZ_Bethe_State& RightState, DP Chem_Pot, stringstream& DAT_outfile) + { + // This function prints the matrix information to the fstream, + // and returns a weighed `data_value' to be multiplied by sumrule_factor, + // to determine if scanning along this thread should be continued. + + // Identify which matrix element is needed from the number of particles in Left and Right states: + + bool fixed_iK = (iKmin == iKmax); + + DP ME = 0.0; + if (!(LeftState.conv && RightState.conv)) ME = 0.0; + + else if (whichDSF == 'Z') + ME = LeftState.E - RightState.E; + else if (whichDSF == 'm') + ME = exp(real(ln_Smin_ME (RightState, LeftState))); + else if (whichDSF == 'z') { + if (LeftState.label == RightState.label) + //MEsq = RightState.chain.Nsites * 0.25 * pow((1.0 - 2.0*RightState.base.Mdown/RightState.chain.Nsites), 2.0); + ME = sqrt(RightState.chain.Nsites * 0.25) * (1.0 - 2.0*RightState.base.Mdown/RightState.chain.Nsites); + else ME = exp(real(ln_Sz_ME (RightState, LeftState))); + } + else if (whichDSF == 'p') + ME = exp(real(ln_Smin_ME (LeftState, RightState))); + else JSCerror("Wrong whichDSF in Compute_Matrix_Element_Contrib."); + + if (is_nan(ME)) ME = 0.0; + + //if (LeftState.dev > 1.0e-16 || RightState.dev > 1.0e-16) ME = 0.0; // kill deviated contributions + + // Do the momentum business: + int iKout = LeftState.iK - RightState.iK; + while(iKout < 0) iKout += RightState.chain.Nsites; + while(iKout >= RightState.chain.Nsites) iKout -= RightState.chain.Nsites; + + // Print information to fstream: + if (iKout >= iKmin && iKout <= iKmax) { + if (whichDSF == 'Z') { + DAT_outfile << endl << setprecision(16) << LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot << "\t" + << iKout << "\t" + //<< LeftState.conv << "\t" + << setprecision(3) << LeftState.dev << "\t" + << LeftState.label; + } + else { + DAT_outfile << endl << setprecision(16) << LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot << "\t" + << iKout << "\t" + << ME << "\t" + //<< LeftState.conv << "\t" + << setprecision(3) << LeftState.dev << "\t" + << LeftState.label; + } + } // if iKmin <= iKout <= iKmax + + // Calculate and return the data_value: + DP data_value = ME * ME; + //DP data_value = (iKout == 0 ? 1.0 : 2.0) * MEsq; + + if (whichDSF == 'Z') // use 1/(1 + omega) + data_value = 1.0/(1.0 + LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot); + else if (fixed_iK) // data value is MEsq * omega: + data_value = ME * ME * (LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot); + + return(data_value); + } + + //DP Compute_Matrix_Element_Contrib (char whichDSF, bool fixed_iK, XXX_Bethe_State& LeftState, + //XXX_Bethe_State& RightState, DP Chem_Pot, fstream& DAT_outfile) + //DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXX_Bethe_State& LeftState, + //XXX_Bethe_State& RightState, DP Chem_Pot, fstream& DAT_outfile) + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXX_Bethe_State& LeftState, + XXX_Bethe_State& RightState, DP Chem_Pot, stringstream& DAT_outfile) + { + // This function prints the matrix element information to the fstream, + // and returns a weighed `data_value' to be multiplied by sumrule_factor, + // to determine if scanning along this thread should be continued. + + // Identify which matrix element is needed from the number of particles in Left and Right states: + + bool fixed_iK = (iKmin == iKmax); + + DP ME = 0.0; + complex ME_CX = 0.0; + + int nrinfrap = 0; // for energy shift from chemical potential + + if (!(LeftState.conv && RightState.conv)) { ME = 0.0; ME_CX = 0.0;} + + else if (whichDSF == 'Z') + ME = LeftState.E - RightState.E; + else if (whichDSF == 'm') + ME = exp(real(ln_Smin_ME (RightState, LeftState))); + else if (whichDSF == 'z') { + // Recognize the presence of an infinite rapidity: + if (LeftState.base.Mdown == RightState.base.Mdown - 1) { // infinite rapidity present, use rescaled S^- matrix element instead of S^z one: + nrinfrap = 1; + // Correction factor for MEsq: Smffsq to Szffsq = 1/(N - 2M + 2) + ME = sqrt(1.0/(RightState.chain.Nsites - 2*RightState.base.Mdown + 2)) * exp(real(ln_Smin_ME (RightState, LeftState))); + } + else { // no infinite rapidity, use S^z matrix element: + if (LeftState.label == RightState.label) + //MEsq = RightState.chain.Nsites * 0.25 * pow((1.0 - 2.0*RightState.base.Mdown/RightState.chain.Nsites), 2.0); + ME = sqrt(RightState.chain.Nsites * 0.25) * (1.0 - 2.0*RightState.base.Mdown/RightState.chain.Nsites); + else ME = exp(real(ln_Sz_ME (RightState, LeftState))); + } + } + else if (whichDSF == 'p') { + // Recognize the presence of two infinite rapidities: + if (LeftState.base.Mdown == RightState.base.Mdown - 1) { // two infinite rapidities, use rescaled S^- matrix element instead of S^+ + nrinfrap = 2; + // Correction factor for MEsq: Smffsq to Spffsq = 2/((N - 2M + 2) (N - 2M + 1)) + ME = sqrt(2.0/((RightState.chain.Nsites - 2*RightState.base.Mdown + 2.0) * (RightState.chain.Nsites - 2*RightState.base.Mdown + 1.0))) + * exp(real(ln_Smin_ME (RightState, LeftState))); + } + else if (LeftState.base.Mdown == RightState.base.Mdown) { // one infinite rapidity, use rescaled S^z matrix element instead of S^+ + nrinfrap = 1; + // Correction factor for MEsq: Szffsq to Spffsq = 4/(N - 2M) + ME = sqrt(4.0/(RightState.chain.Nsites - 2* RightState.base.Mdown)) * exp(real(ln_Sz_ME (RightState, LeftState))); + } + else ME = exp(real(ln_Smin_ME (LeftState, RightState))); + } + else if (whichDSF == 'a') // S^z_j S^z_{j+1} operator + ME = exp(real(ln_Szz_ME (LeftState, RightState))); + else if (whichDSF == 'b') // S^z_j S^-_{j+1} + h.c. operator + ME = exp(real(ln_Szm_p_Smz_ME (RightState, LeftState))); + else if (whichDSF == 'c') // S^-_j S^-{j+1} operator + ME = exp(real(ln_Smm_ME (RightState, LeftState))); + else if (whichDSF == 'q') // Geometric quench + //ME_CX = ln_Overlap (LeftState, RightState); + ME_CX = ln_Overlap (RightState, LeftState); + else JSCerror("Wrong whichDSF in Compute_Matrix_Element_Contrib."); + + if (is_nan(ME)) ME = 0.0; + if (is_nan(norm(ME_CX))) ME_CX = -100.0; + + //if (LeftState.dev > 1.0e-16 || RightState.dev > 1.0e-16) { + //ME = 0.0; ME_CX = (0.0,0.0); // kill deviated contributions + //} + + // Do the momentum business: + int iKout = LeftState.iK - RightState.iK; + while(iKout < 0) iKout += RightState.chain.Nsites; + while(iKout >= RightState.chain.Nsites) iKout -= RightState.chain.Nsites; + + // Print information to fstream: + if (iKout >= iKmin && iKout <= iKmax) { + if (whichDSF == 'Z') { + DAT_outfile << endl << setprecision(16) << LeftState.E - RightState.E - (LeftState.base.Mdown + nrinfrap - RightState.base.Mdown) * Chem_Pot << "\t" + << iKout << "\t" + //<< LeftState.conv << "\t" + << setprecision(3) << LeftState.dev << "\t" + << LeftState.label; + } + else if (whichDSF == 'q') { + DAT_outfile << endl << setprecision(16) << LeftState.E - RightState.E - (LeftState.base.Mdown + nrinfrap - RightState.base.Mdown) * Chem_Pot << "\t" + << iKout << "\t" + << real(ME_CX) << "\t" << imag(ME_CX) - twoPI * int(imag(ME_CX)/twoPI + 1.0e-10) << "\t" + //<< LeftState.conv << "\t" + << setprecision(3) << LeftState.dev << "\t" + << LeftState.label; + } + + else { + DAT_outfile << endl << setprecision(16) << LeftState.E - RightState.E - (LeftState.base.Mdown + nrinfrap - RightState.base.Mdown) * Chem_Pot << "\t" + << iKout << "\t" + << ME << "\t" + //<< LeftState.conv << "\t" + << setprecision(3) << LeftState.dev << "\t" + << LeftState.label; + } + } // if iKmin <= iKout <= iKmax + + // Calculate and return the data_value: + DP data_value = ME * ME; + //DP data_value = (iKout == 0 ? 1.0 : 2.0) * MEsq; + if (whichDSF == 'Z') // use 1/(1 + omega) + data_value = 1.0/(1.0 + LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot); + else if (whichDSF == 'q') + data_value = exp(2.0 * real(ME_CX)); + else if (fixed_iK) // data value is MEsq * omega: + data_value = ME * ME * (LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot); + + return(data_value); + } + + //DP Compute_Matrix_Element_Contrib (char whichDSF, bool fixed_iK, XXZ_gpd_Bethe_State& LeftState, + //XXZ_gpd_Bethe_State& RightState, DP Chem_Pot, fstream& DAT_outfile) + //DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXZ_gpd_Bethe_State& LeftState, + // XXZ_gpd_Bethe_State& RightState, DP Chem_Pot, fstream& DAT_outfile) + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, XXZ_gpd_Bethe_State& LeftState, + XXZ_gpd_Bethe_State& RightState, DP Chem_Pot, stringstream& DAT_outfile) + { + // This function prints the matrix element information to the fstream, + // and returns a weighed `data_value' to be multiplied by sumrule_factor, + // to determine if scanning along this thread should be continued. + + /* + cout << "\t" << LeftState.label << endl << "\t" << LeftState.Ix2 << endl; + cout << "\t0: "; + for (int i = 0; i < LeftState.base.Nrap[0]; ++i) cout << LeftState.lambda[0][i]*2.0/PI << "\t"; + cout << endl; + cout << "\t1: "; + for (int i = 0; i < LeftState.base.Nrap[1]; ++i) cout << LeftState.lambda[1][i]*2.0/PI << "\t"; + cout << endl; + */ + + bool fixed_iK = (iKmin == iKmax); + + + // If any of the rapidities outside of fundamental interval -pi/2 < lambda <= pi/2, set matrix element to zero. + bool rap_in_fundamental = true; + int sum1 = 0; + for (int j = 0; j < LeftState.chain.Nstrings; ++j) { + //for (int alpha = 0; alpha < LeftState.base.Nrap[j]; ++alpha) { + //if (LeftState.lambda[j][alpha] <= -0.5*PI || LeftState.lambda[j][alpha] > 0.5*PI) + //if (LeftState.lambda[j][alpha] <= -0.5*PI || LeftState.lambda[j][alpha] > 0.5*PI) + //rap_in_fundamental = false; + //} + + /* + // TEST 2014 06 26: comment this out, replace by -\pi/2 \leq \lambda \leq \pi/2, see below + if (LeftState.base.Nrap[j] > 0 && LeftState.lambda[j][LeftState.base.Nrap[j] - 1] - LeftState.lambda[j][0] >= PI) + rap_in_fundamental = false; + + sum1 = 0; + for (int k = 0; k < LeftState.chain.Nstrings; ++k) + sum1 += LeftState.base.Nrap[k] * (2 * JSC::min(LeftState.chain.Str_L[j], LeftState.chain.Str_L[k]) - ((j == k) ? 1 : 0)); + // This almost does it: only missing are the states with one on -PI/2 and one on PI/2 + if (LeftState.base.Nrap[j] >= 1 + && (LeftState.Ix2[j][0] <= -(LeftState.chain.Nsites - sum1) + || (LeftState.Ix2[j][LeftState.base.Nrap[j] - 1] - LeftState.Ix2[j][0]) + > 2*(LeftState.chain.Nsites - sum1))) + rap_in_fundamental = false; + */ + + // attempt 2014 06 26 + //for (int alpha = 0; alpha < LeftState.base.Nrap[j]; ++alpha) { + //if (LeftState.lambda[j][alpha] < -0.5*PI || LeftState.lambda[j][alpha] > 0.5*PI) + // rap_in_fundamental = false; + //} + /* + if (LeftState.base.Nrap[j] > 0 && + ((LeftState.lambda[j][LeftState.base.Nrap[j] - 1] - LeftState.lambda[j][0] >= PI) + || LeftState.lambda[j][0] < -0.5*PI + 1.0e-10 + || LeftState.lambda[j][LeftState.base.Nrap[j] - 1] > 0.5*PI + //|| LeftState.lambda[j][0] > 0.5*PI + //((LeftState.lambda[j][LeftState.base.Nrap[j] - 1] - LeftState.lambda[j][0] >= PI - 1.0e-10) + //|| LeftState.lambda[j][0] < -0.5*PI + 1.0e-10 + //|| LeftState.lambda[j][LeftState.base.Nrap[j] - 1] > 0.5*PI + 1.0e-10 + )) // include safety in limits + rap_in_fundamental = false; + */ + /* + if (LeftState.base.Nrap[j] > 0 && + ((LeftState.lambda[j][LeftState.base.Nrap[j] - 1] - LeftState.lambda[j][0] >= PI) + //|| (LeftState.base.Nrap[j] == 1 && fabs(LeftState.lambda[j][0]) > 0.5*PI) + )) + rap_in_fundamental = false; + */ + + // Logic: allow rapidities -PI/2 <= lambda <= PI/2 (including boundaries) + if (LeftState.base.Nrap[j] > 0 && + (LeftState.lambda[j][0] < -PI/2 || LeftState.lambda[j][LeftState.base.Nrap[j] - 1] > PI/2)) + rap_in_fundamental = false; + if (RightState.base.Nrap[j] > 0 && + (RightState.lambda[j][0] < -PI/2 || RightState.lambda[j][RightState.base.Nrap[j] - 1] > PI/2)) + rap_in_fundamental = false; + + // rapidities should also be ordered as the quantum numbers: + for (int alpha = 1; alpha < LeftState.base.Nrap[j]; ++alpha) + if (LeftState.lambda[j][alpha - 1] >= LeftState.lambda[j][alpha]) + rap_in_fundamental = false; + for (int alpha = 1; alpha < RightState.base.Nrap[j]; ++alpha) + if (RightState.lambda[j][alpha - 1] >= RightState.lambda[j][alpha]) + rap_in_fundamental = false; + + } // for int j + + + // Identify which matrix element is needed from the number of particles in Left and Right states: + DP ME = 0.0; + //if (!(LeftState.conv && RightState.conv)) ME = 0.0; + if (!(LeftState.conv && RightState.conv && rap_in_fundamental)) ME = 0.0; + + else if (whichDSF == 'Z') + ME = LeftState.E - RightState.E; + else if (whichDSF == 'm') + ME = exp(real(ln_Smin_ME (RightState, LeftState))); + else if (whichDSF == 'z') { + if (LeftState.label == RightState.label) + //MEsq = RightState.chain.Nsites * 0.25 * pow((1.0 - 2.0*RightState.base.Mdown/RightState.chain.Nsites), 2.0); + ME = sqrt(RightState.chain.Nsites * 0.25) * (1.0 - 2.0*RightState.base.Mdown/RightState.chain.Nsites); + else ME = exp(real(ln_Sz_ME (RightState, LeftState))); + } + else if (whichDSF == 'p') + ME = exp(real(ln_Smin_ME (LeftState, RightState))); + else JSCerror("Wrong whichDSF in Compute_Matrix_Element_Contrib."); + + if (is_nan(ME)) ME = 0.0; + + //if (LeftState.dev > 1.0e+2 || RightState.dev > 1.0e+2) ME = 0.0; // kill deviated contributions + if (fabs(ME) > 1.0) ME = 0.0; + + // Do the momentum business: + int iKout = LeftState.iK - RightState.iK; + while(iKout < 0) iKout += RightState.chain.Nsites; + while(iKout >= RightState.chain.Nsites) iKout -= RightState.chain.Nsites; + + // Print information to fstream: + if (iKout >= iKmin && iKout <= iKmax) { + if (whichDSF == 'Z') { + DAT_outfile << endl << setprecision(16) << LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot << "\t" + << iKout << "\t" + //<< LeftState.conv << "\t" + << setprecision(3) << LeftState.dev << "\t" + << LeftState.label; + } + else { + DAT_outfile << endl << setprecision(16) << LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot << "\t" + << iKout << "\t" + << ME << "\t" + //<< LeftState.conv << "\t" + << setprecision(3) << LeftState.dev << "\t" + << LeftState.label; + + /* + cout << setprecision(16) << LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot << "\t" << iKout << "\t" << ME << "\t" << setprecision(3) << LeftState.dev << "\t" << LeftState.label << "\t" << setprecision(16) << LeftState.lambda[0][0]/PI << "\t" << LeftState.Ix2[0][0] << "\t" << LeftState.lambda[0][LeftState.base.Nrap[0] - 1]/PI << "\t" << LeftState.Ix2[0][LeftState.base.Nrap[0] - 1]; + if (LeftState.base.Nrap[1] > 0) cout << "\t" << LeftState.lambda[1][0]/PI << "\t" << LeftState.Ix2[1][0]; + if (LeftState.lambda[0][0] < -0.5*PI + 1.0e-10 || LeftState.lambda[0][LeftState.base.Nrap[0] - 1] > 0.5*PI - 1.0e-10 || (LeftState.base.Nrap[1] > 0 && (LeftState.lambda[1][0] < -0.5*PI || LeftState.lambda[1][0] > 0.5*PI))) cout << "\t" << "*****"; + cout << endl; + */ + + } + } // if iKmin <= iKout <= iKmax + + // Calculate and return the data_value: + DP data_value = ME * ME; + //DP data_value = (iKout == 0 ? 1.0 : 2.0) * MEsq; + if (whichDSF == 'Z') // use 1/(1 + omega) + data_value = 1.0/(1.0 + LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot); + else if (fixed_iK) // data value is MEsq * omega: + data_value = ME * ME * (LeftState.E - RightState.E - (LeftState.base.Mdown - RightState.base.Mdown) * Chem_Pot); + + return(data_value); + } + + +} // namespace JSC diff --git a/src/HEIS/Heis_Sumrules.cc b/src/HEIS/Heis_Sumrules.cc new file mode 100644 index 0000000..3917d97 --- /dev/null +++ b/src/HEIS/Heis_Sumrules.cc @@ -0,0 +1,255 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/HEIS/Heis_Sumrules.cc + +Purpose: defines sumrule factors for Heisenberg + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + DP X_avg (char xyorz, DP Delta, int N, int M) + { + // Calculates \sum_j < S_j^a S_{j+1}^a >, a = x, y or z. + + DP eps_Delta = 0.00000001; + + if (Delta < 1.0 && 1.0 - Delta < 1.0e-6) eps_Delta *= -1.0; + + // Define the chain: J, Delta, h, Nsites + Heis_Chain chain(1.0, Delta, 0.0, N); + + // Define the base: chain, Mdown + Heis_Base gbase(chain, M); + + // Define the chain: J, Delta, h, Nsites + Heis_Chain chain2(1.0, Delta + eps_Delta, 0.0, N); + + // Define the base: chain, Mdown + Heis_Base gbase2(chain2, M); + + DP E0_Delta = 0.0; + DP E0_Delta_eps = 0.0; + + if (Delta > 0.0 && Delta < 1.0) { + + // Define the ground state + XXZ_Bethe_State gstate(chain, gbase); + + // Compute everything about the ground state + gstate.Compute_All(true); + + E0_Delta = gstate.E; + + // Define the ground state + XXZ_Bethe_State gstate2(chain2, gbase2); + + // Compute everything about the ground state + gstate2.Compute_All(true); + + E0_Delta_eps = gstate2.E; + } + + else if (Delta == 1.0) { + // Define the ground state + XXX_Bethe_State gstate(chain, gbase); + + // Compute everything about the ground state + gstate.Compute_All(true); + + E0_Delta = gstate.E; + + // Define the ground state + XXZ_gpd_Bethe_State gstate2(chain2, gbase2); // need XXZ_gpd here + + // Compute everything about the ground state + gstate2.Compute_All(true); + + E0_Delta_eps = gstate2.E; + } + + else if (Delta > 1.0) { + // Define the ground state + XXZ_gpd_Bethe_State gstate(chain, gbase); + + // Compute everything about the ground state + gstate.Compute_All(true); + + E0_Delta = gstate.E; + + // Define the ground state + XXZ_gpd_Bethe_State gstate2(chain2, gbase2); + + // Compute everything about the ground state + gstate2.Compute_All(true); + + E0_Delta_eps = gstate2.E; + } + + else JSCerror("Wrong anisotropy in S1_sumrule_factor."); + + DP answer = 0.0; + //if (xyorz == 'x' || xyorz == 'y') answer = 0.5 * (E0_Delta - Delta * (E0_Delta_eps - E0_Delta)/eps_Delta); + if (xyorz == 'x' || xyorz == 'y') answer = 0.5 * (E0_Delta - Delta * (E0_Delta_eps - E0_Delta)/eps_Delta); + + // Careful for z ! Hamiltonian defined as S^z S^z - 1/4, so add back N/4: + else if (xyorz == 'z') answer = (E0_Delta_eps - E0_Delta)/eps_Delta + 0.25 * N; + + else JSCerror("option not implemented in X_avg."); + + return(answer); + } + + DP S1_sumrule_factor (char mporz, DP Delta, int N, int M, DP Chem_Pot, int iK) + { + + DP X_x = X_avg ('x', Delta, N, M); + DP X_z = X_avg ('z', Delta, N, M); + + DP sumrule = 0.0; + + if (mporz == 'm' || mporz == 'p') sumrule = - 4.0 * ((1.0 - Delta * cos((twoPI * iK)/N)) * X_x + (Delta - cos((twoPI * iK)/N)) * X_z + 0.5 * Chem_Pot * 0.5 *(N - 2*M))/N; + //if (mporz == 'm' || mporz == 'p') sumrule = - 1.0 * ((1.0 - Delta * cos((twoPI * iK)/N)) * 2.0 * X_x + (Delta - cos((twoPI * iK)/N)) * (X_x + X_z))/N; + + else if (mporz == 'z') sumrule = iK == 0 ? 1.0 : -2.0 * X_x * (1.0 - cos((twoPI * iK)/N))/N; + + else if (mporz == 'Z') sumrule = 1.0; // partition function + else if (mporz == 'a') sumrule = 1.0; + else if (mporz == 'b') sumrule = 1.0; + else if (mporz == 'c') sumrule = 1.0; + + else JSCerror("option not implemented in S1_sumrule_factor."); + + //return(1.0/sumrule); + return(1.0/(sumrule + 1.0e-16)); // sumrule is 0 for iK == 0 or N + } + + DP S1_sumrule_factor (char mporz, DP Delta, int N, int M, DP Chem_Pot, DP X_x, DP X_z, int iK) + { + + //DP X_x = X_avg ('x', Delta, N, M); + //DP X_z = X_avg ('z', Delta, N, M); + + DP sumrule = 0.0; + + if (mporz == 'm' || mporz == 'p') sumrule = - 4.0 * ((1.0 - Delta * cos((twoPI * iK)/N)) * X_x + (Delta - cos((twoPI * iK)/N)) * X_z + 0.5 * Chem_Pot * 0.5 *(N - 2*M))/N; + //if (mporz == 'm' || mporz == 'p') sumrule = - 1.0 * ((1.0 - Delta * cos((twoPI * iK)/N)) * 2.0 * X_x + (Delta - cos((twoPI * iK)/N)) * (X_x + X_z))/N; + + else if (mporz == 'z') sumrule = -2.0 * X_x * (1.0 - cos((twoPI * iK)/N))/N; + + else if (mporz == 'Z') sumrule = 1.0; // partition function + else if (mporz == 'a') sumrule = 1.0; + else if (mporz == 'b') sumrule = 1.0; + else if (mporz == 'c') sumrule = 1.0; + + else JSCerror("option not implemented in S1_sumrule_factor."); + + return(1.0/(sumrule + 1.0e-16)); // sumrule is 0 for iK == 0 or N + } + + + //DP Sumrule_Factor (char whichDSF, Heis_Bethe_State& AveragingState, DP Chem_Pot, bool fixed_iK, int iKneeded) + DP Sumrule_Factor (char whichDSF, Heis_Bethe_State& AveragingState, DP Chem_Pot, int iKmin, int iKmax) + { + DP sumrule_factor = 1.0; + //if (!fixed_iK) { + if (iKmin != iKmax) { + if (whichDSF == 'Z') sumrule_factor = 1.0; + else if (whichDSF == 'm') + sumrule_factor = 1.0/AveragingState.base.Mdown; + else if (whichDSF == 'z') sumrule_factor = 1.0/(0.25 * AveragingState.chain.Nsites); + else if (whichDSF == 'p') sumrule_factor = 1.0/(AveragingState.chain.Nsites - AveragingState.base.Mdown); + else if (whichDSF == 'a') sumrule_factor = 1.0; + else if (whichDSF == 'b') sumrule_factor = 1.0; + else if (whichDSF == 'c') sumrule_factor = 1.0; + else if (whichDSF == 'q') sumrule_factor = 1.0; + + else JSCerror("whichDSF option not consistent in Sumrule_Factor"); + } + //else if (fixed_iK) { + else if (iKmin == iKmax) { + if (whichDSF == 'Z') sumrule_factor = 1.0; + else if (whichDSF == 'm' || whichDSF == 'z' || whichDSF == 'p') + //sumrule_factor = S1_sumrule_factor (whichDSF, AveragingState.chain.Delta, AveragingState.chain.Nsites, AveragingState.base.Mdown, iKneeded); + sumrule_factor = S1_sumrule_factor (whichDSF, AveragingState.chain.Delta, AveragingState.chain.Nsites, AveragingState.base.Mdown, Chem_Pot, iKmax); + else if (whichDSF == 'a') sumrule_factor = 1.0; + else if (whichDSF == 'b') sumrule_factor = 1.0; + else if (whichDSF == 'c') sumrule_factor = 1.0; + else if (whichDSF == 'q') sumrule_factor = 1.0; + + else JSCerror("whichDSF option not consistent in Sumrule_Factor"); + } + + + + return(sumrule_factor); + } + + + void Evaluate_F_Sumrule (string prefix, char whichDSF, const Heis_Bethe_State& AveragingState, DP Chem_Pot, int iKmin_ref, int iKmax_ref) + { + + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + stringstream FSR_stringstream; string FSR_string; + FSR_stringstream << prefix << ".fsr"; + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + + ifstream infile; + infile.open(RAW_Cstr); + if(infile.fail()) JSCerror("Could not open raw input file in Evaluate_F_Sumrule(Heis...)."); + + int iKmin = 0; + int iKmax = AveragingState.chain.Nsites; + int iKmod = AveragingState.chain.Nsites; + + // We run through the data file to chech the f sumrule at each positive momenta: + //Vect Sum_omega_MEsq(0.0, iKmax - iKmin + 1); + Vect Sum_omega_MEsq(0.0, iKmax - iKmin + 1); + + DP omega, ME; + int iK, iKexc; + //int conv; + DP dev; + string label; + + while (infile.peek() != EOF) { + infile >> omega >> iK >> ME >> dev >> label; + //if (iK >= iKmin && iK <= iKmax) Sum_omega_MEsq[iK - iKmin] += omega * ME * ME; + iKexc = iK; + while (iKexc > iKmax && iKexc - iKmod >= iKmin) iKexc -= iKmod; + while (iKexc < iKmin && iKexc + iKmod <= iKmax) iKexc += iKmod; + if (iKexc >= iKmin && iKexc <= iKmax) Sum_omega_MEsq[iKexc - iKmin] += omega * ME * ME; + } + + infile.close(); + + ofstream outfile; + outfile.open(FSR_Cstr); + outfile.precision(16); + + DP X_x = X_avg ('x', AveragingState.chain.Delta, AveragingState.chain.Nsites, AveragingState.base.Mdown); + DP X_z = X_avg ('z', AveragingState.chain.Delta, AveragingState.chain.Nsites, AveragingState.base.Mdown); + + for (int i = iKmin; i <= iKmax; ++i) { + if (i > iKmin) outfile << endl; + outfile << i << "\t" << Sum_omega_MEsq[i] * S1_sumrule_factor (whichDSF, AveragingState.chain.Delta, AveragingState.chain.Nsites, AveragingState.base.Mdown, Chem_Pot, X_x, X_z, i); + } + + outfile.close(); + } + +} // namespace JSC diff --git a/src/HEIS/M_vs_H.cc b/src/HEIS/M_vs_H.cc new file mode 100644 index 0000000..b23979a --- /dev/null +++ b/src/HEIS/M_vs_H.cc @@ -0,0 +1,139 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) 2006-9. + +----------------------------------------------------------- + +File: M_vs_H.cc + +Purpose: field to and from magnetization for Heisenberg + +Last modified: 21/10/09 + + +***********************************************************/ + +#include "JSC.h" + +namespace JSC { + + DP Ezero (DP Delta, int N, int M) + { + // Returns the energy of the ground state with M down spins + + if (M < 0 || M > N/2) JSCerror("M out of bounds in Ezero."); + + DP E = -1.0; // sentinel value + + if (M == 0) E = N * Delta/4.0; + + else { + + Heis_Chain BD1(1.0, Delta, 0.0, N); + + Vect_INT Nrapidities_groundstate(0, BD1.Nstrings); + + Nrapidities_groundstate[0] = M; + + Heis_Base baseconfig_groundstate(BD1, Nrapidities_groundstate); + + if ((Delta > 0.0) && (Delta < 1.0)) { + XXZ_Bethe_State groundstate(BD1, baseconfig_groundstate); + groundstate.Compute_All(true); + E = groundstate.E; + } + + else if (Delta == 1.0) { + XXX_Bethe_State groundstate(BD1, baseconfig_groundstate); + groundstate.Compute_All(true); + E = groundstate.E; + } + + else if (Delta > 1.0) { + XXZ_gpd_Bethe_State groundstate(BD1, baseconfig_groundstate); + groundstate.Compute_All(true); + E = groundstate.E; + } + + else JSCerror("Anisotropy out of bounds in Ezero."); + } + + return(E); + } + + DP H_vs_M (DP Delta, int N, int M) + { + // Assumes dE/dM = 0 = dE_0/dM + h, with dE_0/dM = E_0(M) - E_0 (M - 1) + + DP H = 0.0; + + if (2*M == N) H = 0.0; + + else if (Delta <= 1.0) H = Ezero (Delta, N, M - 1) - Ezero (Delta, N, M); + + return(H); + } + + DP HZmin (DP Delta, int N, int M, Vect_DP& Ezero_ref) + { + if (M < 0 || M > N/2 - 1) { + cout << "M = " << M << endl; + JSCerror("M out of bounds in HZmin."); + } + + if (Ezero_ref[M] == -1.0) Ezero_ref[M] = Ezero(Delta, N, M); + if (Ezero_ref[M + 1] == -1.0) Ezero_ref[M + 1] = Ezero(Delta, N, M + 1); + + return(Ezero_ref[M] - Ezero_ref[M + 1]); + } + + int M_vs_H (DP Delta, int N, DP HZ) + { + // Returns the value of M for given field HZ + + if (HZ < 0.0) JSCerror("Please use a positive field in M_vs_H."); + + else if (HZ == 0.0) return(N/2); + + // Here, -1.0 is a sentinel value. + Vect_DP Ezero(-1.0, N/2 + 1); // contains the GSE[M]. + + // We look for M s.t. HZmin[M] < HZ <= HZmin[M + 1] + + int M_actual = N/4; // start somewhere in middle + int M_step = N/8 - 1; // step + DP HZmin_actual = 0.0; + DP HZmax_actual = 0.0; + bool M_found = false; + + if (HZ >= 1.0 + Delta) M_actual = 0; // saturation + + else { + + HZmin_actual = HZmin (Delta, N, M_actual, Ezero); + HZmax_actual = HZmin (Delta, N, M_actual - 1, Ezero); + + while (!M_found) { + + if (HZmin_actual > HZ) M_actual += M_step; + else if (HZmax_actual <= HZ) M_actual -= M_step; + + M_step = (M_step + 1)/2; + + HZmin_actual = HZmin (Delta, N, M_actual, Ezero); + HZmax_actual = HZmin (Delta, N, M_actual - 1, Ezero); + + M_found = (HZmin_actual < HZ && HZ <= HZmax_actual); + + //cout << "M_actual = " << M_actual << "\tM_step = " << M_step + // << "\tHZmin_actual = " << HZmin_actual << "\tHZmax_actual = " << HZmax_actual << "\tHZ = " << HZ << "\t" << M_found << endl; + } + } + //cout << "M found = " << M_actual << "\tHZmax = " << Ezero[M_actual] - Ezero[M_actual + 1] << "\tHZmin = " << Ezero[M_actual - 1] - Ezero[M_actual] << endl; + + return(M_actual); + } + +} diff --git a/src/HEIS/XXX_Bethe_State.cc b/src/HEIS/XXX_Bethe_State.cc new file mode 100644 index 0000000..b829c7e --- /dev/null +++ b/src/HEIS/XXX_Bethe_State.cc @@ -0,0 +1,626 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/HEIS/XXX_Bethe_State.cc + +Purpose: Defines all functions for XXX_Bethe_State + +******************************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + // Function prototypes + + DP Theta_XXX (DP lambda, int nj, int nk); + DP ddlambda_Theta_XXX (DP lambda, int nj, int nk); + + //*************************************************************************************************** + + // Function definitions: class XXX_Bethe_State + + XXX_Bethe_State::XXX_Bethe_State () + : Heis_Bethe_State() + {}; + + XXX_Bethe_State::XXX_Bethe_State (const XXX_Bethe_State& RefState) // copy constructor + : Heis_Bethe_State(RefState) + { + } + + XXX_Bethe_State::XXX_Bethe_State (const Heis_Chain& RefChain, int M) + : Heis_Bethe_State(RefChain, M) + { + if (RefChain.Delta != 1.0) { + cout << setprecision(16) << RefChain.Delta << endl; + JSCerror("Delta != 1.0 in XXX_Bethe_State constructor"); + } + } + + XXX_Bethe_State::XXX_Bethe_State (const Heis_Chain& RefChain, const Heis_Base& RefBase) + : Heis_Bethe_State(RefChain, RefBase) + { + if (RefChain.Delta != 1.0) { + cout << setprecision(16) << RefChain.Delta << endl; + JSCerror("Delta != 1.0 in XXX_Bethe_State constructor"); + } + } + /* + XXX_Bethe_State::XXX_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref) + : Heis_Bethe_State(RefChain, base_id_ref, type_id_ref) + { + if (RefChain.Delta != 1.0) { + cout << setprecision(16) << RefChain.Delta << endl; + JSCerror("Delta != 1.0 in XXX_Bethe_State constructor"); + } + } + */ + XXX_Bethe_State& XXX_Bethe_State::operator= (const XXX_Bethe_State& RefState) + { + if (this != &RefState) { + chain = RefState.chain; + base = RefState.base; + //offsets = RefState.offsets; + Ix2 = RefState.Ix2; + lambda = RefState.lambda; + BE = RefState.BE; + diffsq = RefState.diffsq; + conv = RefState.conv; + iter = RefState.iter; + iter_Newton = RefState.iter_Newton; + E = RefState.E; + iK = RefState.iK; + K = RefState.K; + lnnorm = RefState.lnnorm; + //base_id = RefState.base_id; + //type_id = RefState.type_id; + //id = RefState.id; + //maxid = RefState.maxid; + //nparticles = RefState.nparticles; + label = RefState.label; + } + return(*this); + } + + // Member functions + + void XXX_Bethe_State::Set_Free_lambdas() + { + // Sets all the rapidities to the solutions of the BAEs without scattering terms + + for (int i = 0; i < chain.Nstrings; ++i) { + + for (int alpha = 0; alpha < base[i]; ++alpha) { + + lambda[i][alpha] = chain.Str_L[i] * 0.5 * tan(PI * 0.5 * Ix2[i][alpha]/chain.Nsites); + + } + } + + return; + } + + bool XXX_Bethe_State::Check_Admissibility(char option) + { + // This function checks the admissibility of the Ix2's of a state: + // returns false if there are higher strings with Ix2 = 0, a totally symmetric distribution of I's at each level, + // and strings of equal length modulo 2 and parity with Ix2 = 0, meaning at least two equal roots in BAE. + + bool answer = true; + //int test1, test3; + Vect Zero_at_level(false, chain.Nstrings); // whether there exists an Ix2 == 0 at a given level + + bool higher_string_on_zero = false; + + for (int j = 0; j < chain.Nstrings; ++j) { + // The following line puts answer to true if there is at least one higher string with zero Ix2 + for (int alpha = 0; alpha < base[j]; ++alpha) if ((Ix2[j][alpha] == 0) && (chain.Str_L[j] > 2) && !(chain.Str_L[j] % 2)) + higher_string_on_zero = true; + for (int alpha = 0; alpha < base[j]; ++alpha) if (Ix2[j][alpha] == 0) Zero_at_level[j] = true; + // NOTE: if base[j] == 0, Zero_at_level[j] remains false. + } + + bool symmetric_state = (*this).Check_Symmetry(); + + bool string_coincidence = false; + // Checks that we have strings of equal length modulo 2 with Ix2 == 0, so equal rapidities, and inadmissibility + for (int j1 = 0; j1 < chain.Nstrings; ++j1) { + for (int j2 = j1 + 1; j2 < chain.Nstrings; ++j2) + if (Zero_at_level[j1] && Zero_at_level[j2] && (!((chain.Str_L[j1] + chain.Str_L[j2])%2))) + string_coincidence = true; + } + /* + bool onep_onem_on_zero = false; + if (option == 'm') { // for Smin, we also exclude symmetric states with 1+ and 1- strings on zero + if (Zero_at_level[0] && Zero_at_level[1]) onep_onem_on_zero = true; + } + */ + answer = !(symmetric_state && (higher_string_on_zero || string_coincidence /*|| onep_onem_on_zero*/)); + + // Now check that no Ix2 is equal to +N (since we take -N into account, and I + N == I by periodicity of exp) + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) if ((Ix2[j][alpha] < -chain.Nsites) || (Ix2[j][alpha] >= chain.Nsites)) answer = false; + + if (!answer) { + E = 0.0; + K = 0.0; + conv = 0; + iter = 0; + iter_Newton = 0; + lnnorm = -100.0; + } + + return(answer); // answer == true: nothing wrong with this Ix2_config + } + + void XXX_Bethe_State::Compute_BE (int j, int alpha) + { + // Fills in the BE members with the value of the Bethe equations. + + DP sumtheta = 0.0; + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += atan(lambda[j][alpha] - lambda[k][beta]); + + else sumtheta += 0.5 * Theta_XXX((lambda[j][alpha] - lambda[k][beta]), chain.Str_L[j], chain.Str_L[k]); + } + sumtheta *= 2.0; + + BE[j][alpha] = 2.0 * atan(2.0 * lambda[j][alpha]/chain.Str_L[j]) - (sumtheta + PI*Ix2[j][alpha])/chain.Nsites; + } + + void XXX_Bethe_State::Compute_BE () + { + // Fills in the BE members with the value of the Bethe equations. + + DP sumtheta = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += atan(lambda[j][alpha] - lambda[k][beta]); + + else sumtheta += 0.5 * Theta_XXX((lambda[j][alpha] - lambda[k][beta]), chain.Str_L[j], chain.Str_L[k]); + } + sumtheta *= 2.0; + + BE[j][alpha] = 2.0 * atan(2.0 * lambda[j][alpha]/chain.Str_L[j]) - (sumtheta + PI*Ix2[j][alpha])/chain.Nsites; + } + } + } + + DP XXX_Bethe_State::Iterate_BAE (int j, int alpha) + { + // Returns a new iteration value for lambda[j][alpha] given BE[j][alpha] + + return(0.5 * chain.Str_L[j] * tan(0.5 * + //(PI * Ix2[j][alpha] + sumtheta)/chain.Nsites + (2.0 * atan(2.0 * lambda[j][alpha]/chain.Str_L[j]) - BE[j][alpha]) + )); + } + + /* + void XXX_Bethe_State::Iterate_BAE () + { + // Recalculates the rapidities by iterating Bethe equations + + Lambda New_lambda(chain, base); + DP sumtheta = 0.0; + DP arg = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += atan(lambda[j][alpha] - lambda[k][beta]); + + else sumtheta += 0.5 * Theta_XXX(lambda[j][alpha] - lambda[k][beta], chain.Str_L[j], chain.Str_L[k]); + } + sumtheta *= 2.0; + + New_lambda[j][alpha] = 0.5 * chain.Str_L[j] * tan((PI * 0.5 * Ix2[j][alpha] + 0.5 * sumtheta)/chain.Nsites); + + } + } + + DP New_diffsq = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + New_diffsq += pow(New_lambda[j][alpha] - lambda[j][alpha], 2.0); + lambda[j][alpha] = 1.0 * New_lambda[j][alpha] + 0.0 * lambda[j][alpha]; + } + } + + diffsq = New_diffsq; + iter++; + + return; + } + */ + /* + void XXX_Bethe_State::Iterate_BAE_Newton () + { + // does one step of a Newton method on the rapidities... + + Vect_DP RHSBAE (0.0, base.Nraptot); // contains RHS of BAEs + Vect_CX dlambda (0.0, base.Nraptot); // contains delta lambda computed from Newton's method + SQMat_CX Gaudin (0.0, base.Nraptot); + Vect_INT indx (base.Nraptot); + DP sumtheta = 0.0; + DP arg = 0.0; + DP fn_arg = 0.0; + DP olddiffsq = diffsq; + + // Compute the RHS of the BAEs: + + int index = 0; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += atan(lambda[j][alpha] - lambda[k][beta]); + + else sumtheta += 0.5 * Theta_XXX((lambda[j][alpha] - lambda[k][beta]), chain.Str_L[j], chain.Str_L[k]); + } + sumtheta *= 2.0; + + RHSBAE[index] = chain.Nsites * 2.0 * atan(2.0 * lambda[j][alpha]/chain.Str_L[j]) - sumtheta - PI*Ix2[j][alpha]; + index++; + } + } + + (*this).Build_Reduced_Gaudin_Matrix (Gaudin); + + for (int i = 0; i < base.Nraptot; ++i) dlambda[i] = - RHSBAE[i]; + + DP d; + ludcmp_CX (Gaudin, indx, d); + lubksb_CX (Gaudin, indx, dlambda); + + diffsq = 0.0; + for (int i = 0; i < base.Nraptot; ++i) diffsq += norm(dlambda[i]); + + // if we've converged, calculate the norm here, since the work has been done... + + if (diffsq < chain.prec) { + lnnorm = 0.0; + for (int j = 0; j < base.Nraptot; j++) lnnorm += log(abs(Gaudin[j][j])); + } + + index = 0; + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + lambda[j][alpha] += real(dlambda[index]); + index++; + } + } + + iter_Newton++; + + return; + } + */ + bool XXX_Bethe_State::Check_Rapidities() + { + bool nonan = true; + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) nonan *= !is_nan(lambda[j][alpha]); + + return nonan; + } + + DP XXX_Bethe_State::String_delta () + { + // Computes the sum of absolute value of \delta^{a, a+1} in string hypothesis, for a given Bethe eigenstate + + DP delta = 0.0; + + int occupied_strings = 0; + for (int i = 0; i < (*this).chain.Nstrings; ++i) if ((*this).chain.Str_L[i] > 1) occupied_strings += (*this).base.Nrap[i]; + + //if ((*this).conv == 0) delta = 1.0; + + if (occupied_strings == 0) delta = 0.0; + + else { + + Vect_DP ln_deltadiff(0.0, 1000); // contains ln |delta^{a, a+1}| + Vect_DP deltadiff(0.0, 1000); // contains |delta^{a, a+1}| + + complex log_BAE_reg = 0.0; + + for (int j = 0; j < (*this).chain.Nstrings; ++j) { + for (int alpha = 0; alpha < (*this).base[j]; ++alpha) { + + ln_deltadiff = 0.0; + + for (int a = 1; a <= (*this).chain.Str_L[j]; ++a) { + + if ((*this).chain.Str_L[j] > 1) { // else the BAE are already 1 + + log_BAE_reg = DP((*this).chain.Nsites) * log(((*this).lambda[j][alpha] + 0.5 * II * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a + 1.0)) + /((*this).lambda[j][alpha] + 0.5 * II * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a - 1.0))); + + for (int k = 0; k < (*this).chain.Nstrings; ++k) + for (int beta = 0; beta < (*this).base[k]; ++beta) + for (int b = 1; b <= (*this).chain.Str_L[k]; ++b) { + if ((j != k) || (alpha != beta) || (a != b - 1)) + + log_BAE_reg += log((*this).lambda[j][alpha] + 0.5 * II * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a ) + - ((*this).lambda[k][beta] + 0.5 * II * ((*this).chain.Str_L[k] + 1.0 - 2.0 * b ) + ) - II ); + + if ((j != k) || (alpha != beta) || (a != b + 1)) + + log_BAE_reg -= log(((*this).lambda[j][alpha] + 0.5 * II * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a ) + ) + - ((*this).lambda[k][beta] + 0.5 * II * ((*this).chain.Str_L[k] + 1.0 - 2.0 * b ) + ) + II ); + } + + // The regular LHS of BAE is now defined. Now sum up the deltas... + + if (a == 1) ln_deltadiff[0] = - real(log_BAE_reg); + + else if (a < (*this).chain.Str_L[j]) ln_deltadiff[a - 1] = ln_deltadiff[a-2] - real(log_BAE_reg); + + else if (a == (*this).chain.Str_L[j]) ln_deltadiff[a-1] = real(log_BAE_reg); + + } // if ((*this).chain.Str_L[j] > 1) + + } // for (int a = 1; ... + + for (int a = 0; a < (*this).chain.Str_L[j]; ++a) { + deltadiff[a] = ln_deltadiff[a] != 0.0 ? exp(ln_deltadiff[a]) : 0.0; + delta += fabs(deltadiff[a]); + } + + } // alpha sum + } // j sum + + if (is_nan(delta)) delta = 1.0; // sentinel + + } // else + + return delta; + } + + + void XXX_Bethe_State::Compute_Energy () + { + DP sum = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum += chain.Str_L[j] / ( 4.0 * lambda[j][alpha] * lambda[j][alpha] + chain.Str_L[j] * chain.Str_L[j]); + } + } + + sum *= - chain.J * 2.0; + + E = sum; + + return; + } + + /* + void XXX_Bethe_State::Compute_Momentum () + { + int sum_Ix2 = 0; + DP sum_M = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + sum_M += base[j]; + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum_Ix2 += Ix2[j][alpha]; + } + } + + iK = (chain.Nsites/2) * int(sum_M) - (sum_Ix2/2); + + while (iK >= chain.Nsites) iK -= chain.Nsites; + while (iK < 0) iK += chain.Nsites; + + K = PI * sum_M - PI * sum_Ix2/chain.Nsites; + + while (K >= 2.0*PI) K -= 2.0*PI; + while (K < 0.0) K += 2.0*PI; + + return; + } + */ + + void XXX_Bethe_State::Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red) + { + + if (Gaudin_Red.size() != base.Nraptot) JSCerror("Passing matrix of wrong size in Build_Reduced_Gaudin_Matrix."); + + int index_jalpha; + int index_kbeta; + + DP sum_hbar_XXX = 0.0; + + index_jalpha = 0; + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + index_kbeta = 0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) { + + if ((j == k) && (alpha == beta)) { + + sum_hbar_XXX = 0.0; + + for (int kp = 0; kp < chain.Nstrings; ++kp) { + for (int betap = 0; betap < base[kp]; ++betap) { + if (!((j == kp) && (alpha == betap))) + sum_hbar_XXX + += ddlambda_Theta_XXX (lambda[j][alpha] - lambda[kp][betap], chain.Str_L[j], chain.Str_L[kp]); + } + } + + Gaudin_Red[index_jalpha][index_kbeta] + = complex ( chain.Nsites * chain.Str_L[j]/(lambda[j][alpha] * lambda[j][alpha] + 0.25 * chain.Str_L[j] * chain.Str_L[j]) + - sum_hbar_XXX); + } + + else { + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + Gaudin_Red[index_jalpha][index_kbeta] = + complex ( 2.0/(pow(lambda[j][alpha] - lambda[k][beta], 2.0) + 1.0)); + + else + Gaudin_Red[index_jalpha][index_kbeta] = + complex (ddlambda_Theta_XXX (lambda[j][alpha] - lambda[k][beta], chain.Str_L[j], chain.Str_L[k])); + } + index_kbeta++; + } + } + index_jalpha++; + } + } + return; + } + + bool XXX_Bethe_State::Check_Finite_rap () + { + bool answer = true; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + if (fabs(lambda[j][alpha]) > 1.0e6) answer = false; + } + } + + return(answer); + } + + // **************************************************************************************************** + + // non-member functions + + DP Theta_XXX (DP lambda, int nj, int nk) + { + DP result; + + if ((nj == 1) && (nk == 1)) result = 2.0 * atan(lambda); + + else { + + result = (nj == nk) ? 0.0 : 2.0 * atan(2.0 * lambda/fabs(nj - nk)); + + for (int a = 1; a < JSC::min(nj, nk); ++a) result += 4.0 * atan(2.0 * lambda/(fabs(nj - nk) + 2*a)); + + result += 2.0 * atan(2.0 * lambda/(nj + nk)); + } + + return (result); + } + + DP ddlambda_Theta_XXX (DP lambda, int nj, int nk) + { + int n = abs(nj - nk); + + DP result = (nj == nk) ? 0.0 : DP(n)/(lambda * lambda + 0.25 * n * n); + + for (int a = 1; a < JSC::min(nj, nk); ++a) result += 2.0 * (n + 2.0*a) + / (lambda * lambda + 0.25 * (n + 2.0*a) * (n + 2.0*a)); + + result += DP(nj + nk)/(lambda * lambda + 0.25 * (nj + nk) * (nj + nk)); + + return (result); + } + /* + DP ddlambda_Theta_XXX (DP lambda, int nj, int nk) + { + DP result = (nj == nk) ? 0.0 : DP(nj - nk)/(lambda * lambda + 0.25 * (nj - nk) * (nj - nk)); + + for (int a = 1; a < JSC::min(nj, nk); ++a) result += 2.0 * (nj - nk + 2.0*a) * (nj - nk + 2.0*a) + / (lambda * lambda + 0.25 * (nj - nk + 2.0*a) * (nj - nk + 2.0*a)); + + result += DP(nj + nk)/(lambda * lambda + 0.25 * (nj + nk) * (nj + nk)); + + return (result); + } + */ + + XXX_Bethe_State Add_Particle_at_Center (const XXX_Bethe_State& RefState) + { + if (2*RefState.base.Mdown == RefState.chain.Nsites) + JSCerror("Trying to add a down spin to a zero-magnetized chain in Add_Particle_at_Center."); + + Vect newM = RefState.base.Nrap; + newM[0] = newM[0] + 1; + + Heis_Base newBase (RefState.chain, newM); + + XXX_Bethe_State ReturnState (RefState.chain, newBase); + + for (int il = 1; il < RefState.chain.Nstrings; ++il) + for (int alpha = 0; alpha < RefState.base.Nrap[il]; ++alpha) + ReturnState.Ix2[il][alpha] = RefState.Ix2[il][alpha]; + + // Add a quantum number in middle (explicitly: to right of index M[0]/2) + // and shift quantum numbers by half-integer away from added one: + ReturnState.Ix2[0][RefState.base.Nrap[0]/2] = RefState.Ix2[0][RefState.base.Nrap[0]/2] - 1; + for (int i = 0; i < RefState.base.Nrap[0] + 1; ++i) + ReturnState.Ix2[0][i + (i >= RefState.base.Nrap[0]/2)] = RefState.Ix2[0][i] - 1 + 2*(i >= RefState.base.Nrap[0]/2); + + return(ReturnState); + } + + + XXX_Bethe_State Remove_Particle_at_Center (const XXX_Bethe_State& RefState) + { + if (RefState.base.Nrap[0] == 0) + JSCerror("Trying to remove a down spin in an empty Nrap[0] state."); + + Vect newM = RefState.base.Nrap; + newM[0] = newM[0] - 1; + + Heis_Base newBase (RefState.chain, newM); + + XXX_Bethe_State ReturnState (RefState.chain, newBase); + + for (int il = 1; il < RefState.chain.Nstrings; ++il) + for (int alpha = 0; alpha < RefState.base.Nrap[il]; ++alpha) + ReturnState.Ix2[il][alpha] = RefState.Ix2[il][alpha]; + + // Remove midmost and shift quantum numbers by half-integer towards removed one: + for (int i = 0; i < RefState.base.Nrap[0]-1; ++i) + ReturnState.Ix2[0][i] = RefState.Ix2[0][i + (i >= RefState.base.Nrap[0]/2)] + 1 - 2*(i >= RefState.base.Nrap[0]/2); + + return(ReturnState); + } + + +} // namespace JSC diff --git a/src/HEIS/XXZ_Bethe_State.cc b/src/HEIS/XXZ_Bethe_State.cc new file mode 100644 index 0000000..78b0ad4 --- /dev/null +++ b/src/HEIS/XXZ_Bethe_State.cc @@ -0,0 +1,646 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/HEIS/XXZ_Bethe_State.cc + +Purpose: Defines all functions for XXZ_Bethe_State + +******************************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + // Function prototypes + + inline DP fbar_XXZ (DP lambda, int par, DP tannzetaover2); + DP Theta_XXZ (DP lambda, int nj, int nk, int parj, int park, DP* tannzetaover2); + DP hbar_XXZ (DP lambda, int n, int par, DP* si_n_anis_over_2); + DP ddlambda_Theta_XXZ (DP lambda, int nj, int nk, int parj, int park, DP* si_n_anis_over_2); + + + //*************************************************************************************************** + + // Function definitions: class XXZ_Bethe_State + + XXZ_Bethe_State::XXZ_Bethe_State () + : Heis_Bethe_State(), sinhlambda(Lambda(chain, 1)), coshlambda(Lambda(chain, 1)), tanhlambda(Lambda(chain, 1)) + {}; + + XXZ_Bethe_State::XXZ_Bethe_State (const XXZ_Bethe_State& RefState) // copy constructor + : Heis_Bethe_State(RefState), sinhlambda(Lambda(RefState.chain, RefState.base)), coshlambda(Lambda(RefState.chain, RefState.base)), + tanhlambda(Lambda(RefState.chain, RefState.base)) + { + // copy arrays into new ones + + //cout << "Calling XXZ state copy constructor." << endl; + for (int j = 0; j < RefState.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < RefState.base[j]; ++j) { + sinhlambda[j][alpha] = RefState.sinhlambda[j][alpha]; + coshlambda[j][alpha] = RefState.coshlambda[j][alpha]; + tanhlambda[j][alpha] = RefState.tanhlambda[j][alpha]; + } + } + //cout << "Done calling XXZ state copy constructor." << endl; + } + + XXZ_Bethe_State::XXZ_Bethe_State (const Heis_Chain& RefChain, int M) + : Heis_Bethe_State(RefChain, M), + sinhlambda(Lambda(RefChain, M)), coshlambda(Lambda(RefChain, M)), tanhlambda(Lambda(RefChain, M)) + { + //cout << "Here in XXZ BS constructor." << endl; + //cout << (*this).lambda[0][0] << endl; + //cout << "OK" << endl; + if ((RefChain.Delta <= -1.0) || (RefChain.Delta >= 1.0)) JSCerror("Delta out of range in XXZ_Bethe_State constructor"); + } + + XXZ_Bethe_State::XXZ_Bethe_State (const Heis_Chain& RefChain, const Heis_Base& RefBase) + : Heis_Bethe_State(RefChain, RefBase), + sinhlambda(Lambda(RefChain, RefBase)), coshlambda(Lambda(RefChain, RefBase)), tanhlambda(Lambda(RefChain, RefBase)) + { + if ((RefChain.Delta <= -1.0) || (RefChain.Delta >= 1.0)) JSCerror("Delta out of range in XXZ_Bethe_State constructor"); + } + /* + XXZ_Bethe_State::XXZ_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref) + : Heis_Bethe_State(RefChain, base_id_ref, type_id_ref), + sinhlambda(Lambda(chain, base)), coshlambda(Lambda(chain, base)), tanhlambda(Lambda(chain, base)) + { + if ((RefChain.Delta <= -1.0) || (RefChain.Delta >= 1.0)) JSCerror("Delta out of range in XXZ_Bethe_State constructor"); + } + */ + + XXZ_Bethe_State& XXZ_Bethe_State::operator= (const XXZ_Bethe_State& RefState) + { + if (this != &RefState) { + chain = RefState.chain; + base = RefState.base; + //offsets = RefState.offsets; + Ix2 = RefState.Ix2; + lambda = RefState.lambda; + BE = RefState.BE; + diffsq = RefState.diffsq; + conv = RefState.conv; + iter = RefState.iter; + iter_Newton = RefState.iter_Newton; + E = RefState.E; + iK = RefState.iK; + K = RefState.K; + lnnorm = RefState.lnnorm; + //base_id = RefState.base_id; + //type_id = RefState.type_id; + //id = RefState.id; + //maxid = RefState.maxid; + //nparticles = RefState.nparticles; + label = RefState.label; + + sinhlambda = RefState.sinhlambda; + coshlambda = RefState.coshlambda; + tanhlambda = RefState.tanhlambda; + } + return(*this); + } + + // Member functions + + void XXZ_Bethe_State::Set_Free_lambdas() + { + // Sets all the rapidities to the solutions of the BAEs without scattering terms + + DP x = 0.0; + for (int i = 0; i < chain.Nstrings; ++i) { + + for (int alpha = 0; alpha < base[i]; ++alpha) { + + if (chain.par[i] == 1) { + //lambda[i][alpha] = atanh(tan(chain.Str_L[i] * 0.5 * chain.anis) * tan(PI * 0.5 * Ix2[i][alpha]/chain.Nsites)); + x = tan(chain.Str_L[i] * 0.5 * chain.anis) * tan(PI * 0.5 * Ix2[i][alpha]/chain.Nsites); + lambda[i][alpha] = atanh(x/sqrt(1.0 + x*x)); // lambda then always initiated real + } + + else if (chain.par[i] == -1) { + //lambda[i][alpha] = atanh(-tan(PI * 0.5 * Ix2[i][alpha]/chain.Nsites)/tan(chain.Str_L[i] * 0.5 * chain.anis)); + x = -tan(PI * 0.5 * Ix2[i][alpha]/chain.Nsites)/tan(chain.Str_L[i] * 0.5 * chain.anis); + lambda[i][alpha] = atanh(x/sqrt(1.0 + x*x)); // lambda then always initiated real + } + + else JSCerror("Invalid parities in Set_Free_lambdas."); + //cout << tan(chain.Str_L[i] * 0.5 * chain.anis) << endl; + //cout << "Set_Free_lambdas: " << i << "\t" << alpha << "\t" << lambda[i][alpha] << "\t" << tan(chain.Str_L[i] * 0.5 * chain.anis) * tan(PI * 0.5 * Ix2[i][alpha]/chain.Nsites) << endl; + + } + } + + return; + } + + void XXZ_Bethe_State::Compute_sinhlambda() + { + for (int j = 0; j < chain.Nstrings; ++j) { + + for (int alpha = 0; alpha < base[j]; ++alpha) sinhlambda[j][alpha] = sinh(lambda[j][alpha]); + } + return; + } + + void XXZ_Bethe_State::Compute_coshlambda() + { + for (int j = 0; j < chain.Nstrings; ++j) { + + for (int alpha = 0; alpha < base[j]; ++alpha) coshlambda[j][alpha] = cosh(lambda[j][alpha]); + } + return; + } + + void XXZ_Bethe_State::Compute_tanhlambda() + { + for (int j = 0; j < chain.Nstrings; ++j) { + + for (int alpha = 0; alpha < base[j]; ++alpha) tanhlambda[j][alpha] = tanh(lambda[j][alpha]); + } + return; + } + + bool XXZ_Bethe_State::Check_Admissibility(char option) + { + // This function checks the admissibility of the Ix2's of a state: + // returns false if there are higher strings with Ix2 = 0, a totally symmetric distribution of I's at each level, + // and strings of equal length modulo 2 and parity with Ix2 = 0, meaning at least two equal roots in BAE. + + bool answer = true; + Vect Zero_at_level(false, chain.Nstrings); // whether there exists an Ix2 == 0 at a given level + + bool higher_string_on_zero = false; + for (int j = 0; j < chain.Nstrings; ++j) { + // The following line puts answer to true if there is at least one higher string with zero Ix2 + for (int alpha = 0; alpha < base[j]; ++alpha) if ((Ix2[j][alpha] == 0) && (chain.Str_L[j] >= 2) /*&& !(chain.Str_L[j] % 2)*/) + higher_string_on_zero = true; + for (int alpha = 0; alpha < base[j]; ++alpha) if (Ix2[j][alpha] == 0) Zero_at_level[j] = true; + // NOTE: if base[j] == 0, Zero_at_level[j] remains false. + } + + // check symmetry of Ix2 at each level, if there exists a potentially risky Ix2... + + bool symmetric_state = (*this).Check_Symmetry(); + + bool string_coincidence = false; + for (int j1 = 0; j1 < chain.Nstrings; ++j1) { + for (int j2 = j1 + 1; j2 < chain.Nstrings; ++j2) + if (Zero_at_level[j1] && Zero_at_level[j2] && (chain.par[j1] == chain.par[j2]) && (!((chain.Str_L[j1] + chain.Str_L[j2])%2))) + string_coincidence = true; + } + + bool M_odd_and_onep_on_zero = false; + if (option == 'z') { // for Sz, if M is odd, exclude symmetric states with a 1+ on zero + // (zero rapidities in left and right states, so FF det not defined). + bool is_ground_state = base.Nrap[0] == base.Mdown && Ix2[0][0] == -(base.Mdown - 1) && Ix2[0][base.Mdown-1] == base.Mdown - 1; + if (Zero_at_level[0] && (base.Mdown % 2) && !is_ground_state) M_odd_and_onep_on_zero = true; + } + + bool onep_onem_on_zero = false; + if (option == 'm' || option == 'p') { // for Smin, we also exclude symmetric states with 1+ and 1- strings on zero + if (Zero_at_level[0] && Zero_at_level[1]) onep_onem_on_zero = true; + } + + answer = !(symmetric_state && (higher_string_on_zero || string_coincidence || onep_onem_on_zero || M_odd_and_onep_on_zero)); + + // Now check that no Ix2 is equal to +N (since we take -N into account, and I + N == I by periodicity of exp) + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) if ((Ix2[j][alpha] < -chain.Nsites) || (Ix2[j][alpha] >= chain.Nsites)) answer = false; + + if (!answer) { + E = 0.0; + K = 0.0; + conv = 0; + iter = 0; + iter_Newton = 0; + lnnorm = -100.0; + } + + return(answer); // answer == true: nothing wrong with this Ix2_config + } + + void XXZ_Bethe_State::Compute_BE (int j, int alpha) + { + tanhlambda[j][alpha] = tanh(lambda[j][alpha]); + + DP sumtheta = 0.0; + + for (int k = 0; k < chain.Nstrings; ++k) + for (int beta = 0; beta < base[k]; ++beta) { + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += (chain.par[j] == chain.par[k]) + ? atan((tanhlambda[j][alpha] - tanhlambda[k][beta])/((1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta]) * chain.ta_n_anis_over_2[2])) + : - atan(((tanhlambda[j][alpha] - tanhlambda[k][beta])/(1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta])) * chain.ta_n_anis_over_2[2]) ; + else sumtheta += 0.5 * Theta_XXZ((tanhlambda[j][alpha] - tanhlambda[k][beta])/(1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.par[j], chain.par[k], chain.ta_n_anis_over_2); + } + sumtheta *= 2.0; + + BE[j][alpha] = ((chain.par[j] == 1) ? 2.0 * atan(tanhlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) + : -2.0 * atan(tanhlambda[j][alpha] * chain.ta_n_anis_over_2[chain.Str_L[j]])) - (sumtheta + PI*Ix2[j][alpha])/chain.Nsites; + + } + + void XXZ_Bethe_State::Compute_BE () + { + // Fills in the BE members with the value of the Bethe equations. + + (*this).Compute_tanhlambda(); + + DP sumtheta = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) + for (int beta = 0; beta < base[k]; ++beta) { + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += (chain.par[j] == chain.par[k]) + ? atan((tanhlambda[j][alpha] - tanhlambda[k][beta])/((1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta]) * chain.ta_n_anis_over_2[2])) + : - atan(((tanhlambda[j][alpha] - tanhlambda[k][beta])/(1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta])) * chain.ta_n_anis_over_2[2]) ; + else sumtheta += 0.5 * Theta_XXZ((tanhlambda[j][alpha] - tanhlambda[k][beta])/(1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.par[j], chain.par[k], chain.ta_n_anis_over_2); + } + sumtheta *= 2.0; + + BE[j][alpha] = ((chain.par[j] == 1) ? 2.0 * atan(tanhlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) + : -2.0 * atan(tanhlambda[j][alpha] * chain.ta_n_anis_over_2[chain.Str_L[j]])) - (sumtheta + PI*Ix2[j][alpha])/chain.Nsites; + + //if (is_nan(BE[j][alpha])) cout << "BE nan: " << j << "\t" << alpha << "\t" << lambda[j][alpha] << "\t" << tanhlambda[j][alpha] << endl; + } + } + + DP XXZ_Bethe_State::Iterate_BAE (int j, int alpha) + { + // Returns a new iteration value for lambda[j][alpha] given tanhlambda and BE Lambdas + // Assumes that tanhlambda[][] and BE[][] have been computed. + + DP new_lambda = 0.0; + DP arg = 0.0; + + if (chain.par[j] == 1) arg = chain.ta_n_anis_over_2[chain.Str_L[j]] + * tan(0.5 * + //(PI * Ix2[j][alpha] + sumtheta)/chain.Nsites + (2.0 * atan(tanhlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) - BE[j][alpha]) + ); + + else if (chain.par[j] == -1) arg = -tan(0.5 * + //(PI * Ix2[j][alpha] + sumtheta)/chain.Nsites) + (-2.0 * atan(tanhlambda[j][alpha] * chain.ta_n_anis_over_2[chain.Str_L[j]]) - BE[j][alpha])) + /chain.ta_n_anis_over_2[chain.Str_L[j]]; + + if (fabs(arg) < 1.0) { + new_lambda = atanh(arg); + } + + else { + new_lambda = lambda[j][alpha]; // back to drawing board... + int block = 0; // counter to prevent runaway while loop + DP new_tanhlambda = 0.0; + DP sumtheta = 0.0; + arg = 10.0; // reset value to start while loop + while ((fabs(arg) > 1.0) && (block++ < 100)) { // recompute the diverging root on its own... + new_lambda *= 1.01; // try to go slowly towards infinity... + new_tanhlambda = tanh(new_lambda); + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += (chain.par[j] == chain.par[k]) + ? atan((new_tanhlambda - tanhlambda[k][beta])/((1.0 - new_tanhlambda * tanhlambda[k][beta]) * chain.ta_n_anis_over_2[2])) + : - atan(((new_tanhlambda - tanhlambda[k][beta])/(1.0 - new_tanhlambda * tanhlambda[k][beta])) * chain.ta_n_anis_over_2[2]) ; + else sumtheta += 0.5 * Theta_XXZ((new_tanhlambda - tanhlambda[k][beta])/(1.0 - new_tanhlambda * tanhlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.par[j], chain.par[k], chain.ta_n_anis_over_2); + } + sumtheta *= 2.0; + if (chain.par[j] == 1) arg = chain.ta_n_anis_over_2[chain.Str_L[j]] * tan(0.5 * (PI * Ix2[j][alpha] + sumtheta)/chain.Nsites); + + else if (chain.par[j] == -1) arg = -tan(0.5 * (PI * Ix2[j][alpha] + sumtheta)/chain.Nsites)/chain.ta_n_anis_over_2[chain.Str_L[j]]; + + else JSCerror("Invalid parities in Iterate_BAE."); + + } + + if (fabs(arg) < 1.0) { + new_lambda = atanh(arg); + } + + //else cout << "Rapidity blows up !\t" << lambda[j][alpha] << "\t" << new_lambda << endl; + } // else + + return(new_lambda); + } + + bool XXZ_Bethe_State::Check_Rapidities() + { + bool nonan = true; + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) nonan *= !is_nan(lambda[j][alpha]); + + return nonan; + } + + DP XXZ_Bethe_State::String_delta () + { + // Computes the sum of absolute value of \delta^{a, a+1} in string hypothesis, for a given bethe eigenstate + + DP delta = 0.0; + + int occupied_strings = 0; + for (int i = 0; i < (*this).chain.Nstrings; ++i) if ((*this).chain.Str_L[i] > 1) occupied_strings += (*this).base.Nrap[i]; + + //if ((*this).conv == 0) delta = 1.0; + + if (occupied_strings == 0) delta = 0.0; + + else { + + Vect_DP ln_deltadiff(0.0, 1000); // contains ln |delta^{a, a+1}| + Vect_DP deltadiff(0.0, 1000); // contains |delta^{a, a+1}| + + complex log_BAE_reg = 0.0; + + for (int j = 0; j < (*this).chain.Nstrings; ++j) { + for (int alpha = 0; alpha < (*this).base[j]; ++alpha) { + + ln_deltadiff = 0.0; + + for (int a = 1; a <= (*this).chain.Str_L[j]; ++a) { + + if ((*this).chain.Str_L[j] > 1) { // else the BAE are already 1 + + log_BAE_reg = DP((*this).chain.Nsites) * log(sinh((*this).lambda[j][alpha] + + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a + 1.0) + + 0.25 * II * PI * (1.0 - (*this).chain.par[j])) + /sinh((*this).lambda[j][alpha] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - (*this).chain.par[j]))); + + for (int k = 0; k < (*this).chain.Nstrings; ++k) + for (int beta = 0; beta < (*this).base[k]; ++beta) + for (int b = 1; b <= (*this).chain.Str_L[k]; ++b) { + if ((j != k) || (alpha != beta) || (a != b - 1)) + + log_BAE_reg += log(sinh(((*this).lambda[j][alpha] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a ) + + 0.25 * II * PI * (1.0 - (*this).chain.par[j])) + - ((*this).lambda[k][beta] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[k] + 1.0 - 2.0 * b ) + + 0.25 * II * PI * (1.0 - (*this).chain.par[k])) - II * (*this).chain.anis)); + + if ((j != k) || (alpha != beta) || (a != b + 1)) + + log_BAE_reg -= log(sinh(((*this).lambda[j][alpha] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a ) + + 0.25 * II * PI * (1.0 - (*this).chain.par[j])) + - ((*this).lambda[k][beta] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[k] + 1.0 - 2.0 * b ) + + 0.25 * II * PI * (1.0 - (*this).chain.par[k])) + II * (*this).chain.anis)); + } + + // The regular LHS of BAE is now defined. Now sum up the deltas... + + if (a == 1) ln_deltadiff[0] = - real(log_BAE_reg); + + else if (a < (*this).chain.Str_L[j]) ln_deltadiff[a - 1] = ln_deltadiff[a-2] - real(log_BAE_reg); + + else if (a == (*this).chain.Str_L[j]) ln_deltadiff[a-1] = real(log_BAE_reg); + + } // if ((*this).chain.Str_L[j] > 1) + + } // for (int a = 1; ... + + for (int a = 0; a < (*this).chain.Str_L[j]; ++a) { + deltadiff[a] = ln_deltadiff[a] != 0.0 ? exp(ln_deltadiff[a]) : 0.0; + delta += fabs(deltadiff[a]); + } + + } // alpha sum + } // j sum + + if (is_nan(delta)) delta = 1.0; // sentinel + + } // else + + return delta; + } + + + void XXZ_Bethe_State::Compute_Energy () + { + DP sum = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum += sin(chain.Str_L[j] * chain.anis) / (chain.par[j] * cosh(2.0 * lambda[j][alpha]) - cos(chain.Str_L[j] * chain.anis)); + } + } + + sum *= - chain.J * sin(chain.anis); + + E = sum; + + return; + } + + /* + void XXZ_Bethe_State::Compute_Momentum () + { + int sum_Ix2 = 0; + DP sum_M = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + sum_M += 0.5 * (1.0 + chain.par[j]) * base[j]; + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum_Ix2 += Ix2[j][alpha]; + } + } + + iK = (chain.Nsites/2) * int(sum_M + 0.1) - (sum_Ix2/2); // + 0.1: for safety... + + while (iK >= chain.Nsites) iK -= chain.Nsites; + while (iK < 0) iK += chain.Nsites; + + K = PI * sum_M - PI * sum_Ix2/chain.Nsites; + + while (K >= 2.0*PI) K -= 2.0*PI; + while (K < 0.0) K += 2.0*PI; + + return; + } + */ + void XXZ_Bethe_State::Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red) + { + + if (Gaudin_Red.size() != base.Nraptot) JSCerror("Passing matrix of wrong size in Build_Reduced_Gaudin_Matrix."); + + int index_jalpha; + int index_kbeta; + + DP sum_hbar_XXZ = 0.0; + + DP sinzetasq = pow(sin(chain.anis), 2.0); + + (*this).Compute_sinhlambda(); + (*this).Compute_coshlambda(); + + index_jalpha = 0; + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + index_kbeta = 0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) { + + if ((j == k) && (alpha == beta)) { + + sum_hbar_XXZ = 0.0; + + for (int kp = 0; kp < chain.Nstrings; ++kp) { + for (int betap = 0; betap < base[kp]; ++betap) { + if (!((j == kp) && (alpha == betap))) + sum_hbar_XXZ + += ddlambda_Theta_XXZ (lambda[j][alpha] - lambda[kp][betap], chain.Str_L[j], chain.Str_L[kp], chain.par[j], chain.par[kp], + chain.si_n_anis_over_2); + } + } + + Gaudin_Red[index_jalpha][index_kbeta] + = complex ( chain.Nsites * hbar_XXZ (lambda[j][alpha], chain.Str_L[j], chain.par[j], chain.si_n_anis_over_2) - sum_hbar_XXZ); + } + + else { + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + Gaudin_Red[index_jalpha][index_kbeta] = + complex ((chain.par[j] * chain.par[k] == 1) + ? chain.si_n_anis_over_2[4]/(pow(sinhlambda[j][alpha] * coshlambda[k][beta] + - coshlambda[j][alpha] * sinhlambda[k][beta], 2.0) + sinzetasq) + : chain.si_n_anis_over_2[4]/(-pow(coshlambda[j][alpha] * coshlambda[k][beta] + - sinhlambda[j][alpha] * sinhlambda[k][beta], 2.0) + sinzetasq) ); + else + Gaudin_Red[index_jalpha][index_kbeta] = complex (ddlambda_Theta_XXZ (lambda[j][alpha] - lambda[k][beta], chain.Str_L[j], chain.Str_L[k], + chain.par[j], chain.par[k], chain.si_n_anis_over_2)); + } + index_kbeta++; + } + } + index_jalpha++; + } + } + return; + } + + // **************************************************************************************************** + + // non-member functions + + inline DP fbar_XXZ (DP tanhlambda, int par, DP tannzetaover2) + { + DP result = 0.0; + + if (par == 1) result = 2.0 * atan(tanhlambda/tannzetaover2); + + else if (par == -1) result = -2.0 * atan(tanhlambda * tannzetaover2); + + else JSCerror("Faulty parity in fbar_XXZ."); + + return (result); + } + + DP Theta_XXZ (DP tanhlambda, int nj, int nk, int parj, int park, DP* tannzetaover2) + { + DP result = 0.0; + + if ((nj == 1) && (nk == 1)) result = fbar_XXZ(tanhlambda, parj*park, tannzetaover2[2]); + + else { + + result = (nj == nk) ? 0.0 : fbar_XXZ(tanhlambda, parj*park, tannzetaover2[fabs(nj - nk)]); + + for (int a = 1; a < JSC::min(nj, nk); ++a) result += 2.0 * fbar_XXZ(tanhlambda, parj*park, tannzetaover2[fabs(nj - nk) + 2*a]); + + result += fbar_XXZ(tanhlambda, parj*park, tannzetaover2[nj + nk]); + } + + return (result); + } + + DP hbar_XXZ (DP lambda, int n, int par, DP* si_n_anis_over_2) + { + DP result = 0.0; + + if (par == 1) result = si_n_anis_over_2[2*n]/(pow(sinh(lambda), 2.0) + pow(si_n_anis_over_2[n], 2.0)); + + else if (par == -1) result = si_n_anis_over_2[2*n]/(-pow(cosh(lambda), 2.0) + pow(si_n_anis_over_2[n], 2.0)); + + else JSCerror("Faulty parity in hbar_XXZ."); + + return (result); + } + + DP ddlambda_Theta_XXZ (DP lambda, int nj, int nk, int parj, int park, DP* si_n_anis_over_2) + { + DP result = (nj == nk) ? 0.0 : hbar_XXZ(lambda, fabs(nj - nk), parj*park, si_n_anis_over_2); + + for (int a = 1; a < JSC::min(nj, nk); ++a) result += 2.0 * hbar_XXZ(lambda, fabs(nj - nk) + 2*a, parj*park, si_n_anis_over_2); + + result += hbar_XXZ(lambda, nj + nk, parj*park, si_n_anis_over_2); + + return (result); + } + + + XXZ_Bethe_State Add_Particle_at_Center (const XXZ_Bethe_State& RefState) + { + if (2*RefState.base.Mdown == RefState.chain.Nsites) + JSCerror("Trying to add a down spin to a zero-magnetized chain in Add_Particle_at_Center."); + + Vect newM = RefState.base.Nrap; + newM[0] = newM[0] + 1; + + Heis_Base newBase (RefState.chain, newM); + + XXZ_Bethe_State ReturnState (RefState.chain, newBase); + + for (int il = 1; il < RefState.chain.Nstrings; ++il) + for (int alpha = 0; alpha < RefState.base.Nrap[il]; ++alpha) + ReturnState.Ix2[il][alpha] = RefState.Ix2[il][alpha]; + + // Add a quantum number in middle (explicitly: to right of index M[0]/2) + // and shift quantum numbers by half-integer away from added one: + ReturnState.Ix2[0][RefState.base.Nrap[0]/2] = RefState.Ix2[0][RefState.base.Nrap[0]/2] - 1; + for (int i = 0; i < RefState.base.Nrap[0] + 1; ++i) + ReturnState.Ix2[0][i + (i >= RefState.base.Nrap[0]/2)] = RefState.Ix2[0][i] - 1 + 2*(i >= RefState.base.Nrap[0]/2); + + return(ReturnState); + } + + + XXZ_Bethe_State Remove_Particle_at_Center (const XXZ_Bethe_State& RefState) + { + if (RefState.base.Nrap[0] == 0) + JSCerror("Trying to remove a down spin in an empty Nrap[0] state."); + + Vect newM = RefState.base.Nrap; + newM[0] = newM[0] - 1; + + Heis_Base newBase (RefState.chain, newM); + + XXZ_Bethe_State ReturnState (RefState.chain, newBase); + + for (int il = 1; il < RefState.chain.Nstrings; ++il) + for (int alpha = 0; alpha < RefState.base.Nrap[il]; ++alpha) + ReturnState.Ix2[il][alpha] = RefState.Ix2[il][alpha]; + + // Remove midmost and shift quantum numbers by half-integer towards removed one: + for (int i = 0; i < RefState.base.Nrap[0]-1; ++i) + ReturnState.Ix2[0][i] = RefState.Ix2[0][i + (i >= RefState.base.Nrap[0]/2)] + 1 - 2*(i >= RefState.base.Nrap[0]/2); + + return(ReturnState); + } + +} // namespace JSC diff --git a/src/HEIS/XXZ_gpd_Bethe_State.cc b/src/HEIS/XXZ_gpd_Bethe_State.cc new file mode 100644 index 0000000..02577e8 --- /dev/null +++ b/src/HEIS/XXZ_gpd_Bethe_State.cc @@ -0,0 +1,814 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/HEIS/XXZ_gpd_Bethe_State.cc + +Purpose: Defines all functions for XXZ_gpd_Bethe_State + +******************************************************************/ + + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + // Function prototypes + + inline DP fbar_XXZ_gpd (DP lambda, int par, DP tanhnetaover2); + DP Theta_XXZ_gpd (DP lambda, int nj, int nk, DP* tanhnetaover2); + DP hbar_XXZ_gpd (DP lambda, int n, DP* si_n_anis_over_2); + DP ddlambda_Theta_XXZ_gpd (DP lambda, int nj, int nk, DP* si_n_anis_over_2); + + + //*************************************************************************************************** + + // Function definitions: class XXZ_gpd_Bethe_State + + XXZ_gpd_Bethe_State::XXZ_gpd_Bethe_State () + : Heis_Bethe_State(), sinlambda(Lambda(chain, 1)), coslambda(Lambda(chain, 1)), tanlambda(Lambda(chain, 1)) + {}; + + XXZ_gpd_Bethe_State::XXZ_gpd_Bethe_State (const XXZ_gpd_Bethe_State& RefState) // copy constructor + : Heis_Bethe_State(RefState), + sinlambda(Lambda(RefState.chain, RefState.base)), coslambda(Lambda(RefState.chain, RefState.base)), + tanlambda(Lambda(RefState.chain, RefState.base)) + { + // copy arrays into new ones + + for (int j = 0; j < RefState.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < RefState.base[j]; ++j) { + sinlambda[j][alpha] = RefState.sinlambda[j][alpha]; + coslambda[j][alpha] = RefState.coslambda[j][alpha]; + tanlambda[j][alpha] = RefState.tanlambda[j][alpha]; + } + } + } + + XXZ_gpd_Bethe_State::XXZ_gpd_Bethe_State (const Heis_Chain& RefChain, int M) + : Heis_Bethe_State(RefChain, M), + sinlambda(Lambda(RefChain, M)), coslambda(Lambda(RefChain, M)), tanlambda(Lambda(RefChain, M)) + { + if (RefChain.Delta <= 1.0) JSCerror("Delta too low in XXZ_gpd_Bethe_State constructor"); + } + + XXZ_gpd_Bethe_State::XXZ_gpd_Bethe_State (const Heis_Chain& RefChain, const Heis_Base& RefBase) + : Heis_Bethe_State(RefChain, RefBase), + sinlambda(Lambda(RefChain, RefBase)), coslambda(Lambda(RefChain, RefBase)), + tanlambda(Lambda(RefChain, RefBase)) + { + if (RefChain.Delta <= 1.0) JSCerror("Delta too low in XXZ_gpd_Bethe_State constructor"); + } + /* + XXZ_gpd_Bethe_State::XXZ_gpd_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref) + : Heis_Bethe_State(RefChain, base_id_ref, type_id_ref), + sinlambda(Lambda(chain, base)), coslambda(Lambda(chain, base)), tanlambda(Lambda(chain, base)) + { + if (RefChain.Delta <= 1.0) JSCerror("Delta too low in XXZ_gpd_Bethe_State constructor"); + } + */ + XXZ_gpd_Bethe_State& XXZ_gpd_Bethe_State::operator= (const XXZ_gpd_Bethe_State& RefState) + { + if (this != &RefState) { + chain = RefState.chain; + base = RefState.base; + //offsets = RefState.offsets; + Ix2 = RefState.Ix2; + lambda = RefState.lambda; + BE = RefState.BE; + diffsq = RefState.diffsq; + conv = RefState.conv; + iter = RefState.iter; + iter_Newton = RefState.iter_Newton; + E = RefState.E; + iK = RefState.iK; + K = RefState.K; + lnnorm = RefState.lnnorm; + //base_id = RefState.base_id; + //type_id = RefState.type_id; + //id = RefState.id; + //maxid = RefState.maxid; + //nparticles = RefState.nparticles; + label = RefState.label; + + sinlambda = RefState.sinlambda; + coslambda = RefState.coslambda; + tanlambda = RefState.tanlambda; + } + return(*this); + } + + + // Member functions + + void XXZ_gpd_Bethe_State::Set_Free_lambdas() + { + // Sets all the rapidities to the solutions of the BAEs without scattering terms + + for (int i = 0; i < chain.Nstrings; ++i) { + + for (int alpha = 0; alpha < base[i]; ++alpha) { + + lambda[i][alpha] = atan((tanh(chain.Str_L[i] * 0.5 * chain.anis) * tan(PI * 0.5 * Ix2[i][alpha]/chain.Nsites))); + + } + } + + return; + } + + void XXZ_gpd_Bethe_State::Compute_sinlambda() + { + for (int j = 0; j < chain.Nstrings; ++j) { + + for (int alpha = 0; alpha < base[j]; ++alpha) sinlambda[j][alpha] = sin(lambda[j][alpha]); + } + return; + } + + void XXZ_gpd_Bethe_State::Compute_coslambda() + { + for (int j = 0; j < chain.Nstrings; ++j) { + + for (int alpha = 0; alpha < base[j]; ++alpha) coslambda[j][alpha] = cos(lambda[j][alpha]); + } + return; + } + + void XXZ_gpd_Bethe_State::Compute_tanlambda() + { + for (int j = 0; j < chain.Nstrings; ++j) { + + for (int alpha = 0; alpha < base[j]; ++alpha) { + tanlambda[j][alpha] = tan(lambda[j][alpha]); + //if (lambda[j][alpha] > 0.5*PI) cout << "Rapidity higher than 0.5*PI: j = " << j << "\talpha = " << alpha << "\trap = " << lambda[j][alpha] << "\ttan = " << tanlambda[j][alpha] << endl; + } + } + return; + } + + bool XXZ_gpd_Bethe_State::Check_Admissibility(char option) + { + // This function checks the admissibility of the Ix2's of a state: + // returns false if there are higher strings with Ix2 = 0, a totally symmetric distribution of I's at each level, + // and strings of equal length modulo 2 and parity with Ix2 = 0, meaning at least two equal roots in BAE. + + bool answer = true; + Vect Zero_at_level(false, chain.Nstrings); // whether there exists an Ix2 == 0 at a given level + + /* + Vect min_Ix2_max_busy(false, chain.Nstrings); + Vect plus_Ix2_max_busy(false, chain.Nstrings); + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + if (Ix2[j][alpha] == -base.Ix2_max[j]) min_Ix2_max_busy[j] = true; + if (Ix2[j][alpha] == base.Ix2_max[j]) plus_Ix2_max_busy[j] = true; + } + */ + /* + // State is not admissible if this is false: -N/2 + 1 \leq \sum I^j_{\alpha} \leq N + int sum_all_Ix2 = 0; + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum_all_Ix2 += Ix2[j][alpha]; + } + if (sum_all_Ix2 > chain.Nsites || sum_all_Ix2 <= -chain.Nsites) { + cout << "\tSum Ix2 out of fundamental interval: sum_all_Ix2 = " << sum_all_Ix2 << "\tfor N = " << chain.Nsites << endl; + return(false); + } + */ + //Deactivated 2014 06 11, put in Heis_Form_Factor_Entry.cc + /* + // State is not admissible if I_max > 1/2 (N - \sum_k t_{jk} M_k) or I_min < -1/2 (N - sum_k...) + 1 at any level: + int sum1 = 0; + for (int j = 0; j < chain.Nstrings; ++j) { + sum1 = 0; + for (int k = 0; k < chain.Nstrings; ++k) { + sum1 += base[k] * (2 * JSC::min(chain.Str_L[j], chain.Str_L[k]) - ((j == k) ? 1 : 0)); + } + // Define limits... + //if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; + + // This almost does it: only missing are the states with one on -PI/2 and one on PI/2 + if (base[j] >= 1 && (Ix2[j][0] <= -(chain.Nsites - sum1) || + (Ix2[j][base[j] - 1] - Ix2[j][0]) > 2*(chain.Nsites - sum1))) { + //cout << "\tAn Ix2 is out of interval at level " << j << endl; + //cout << Ix2[j][base[j] - 1] << "\t" << Ix2[j][0] << "\t" << chain.Nsites << "\t" << sum1 << endl; + return(false); + } + } + */ + + /* + // State is not admissible if both a min_ and plus_Ix2_max are busy simultaneously: + bool is_a_min_Ix2_max_busy = false; + for (int j = 0; j < chain.Nstrings; ++j) if (min_Ix2_max_busy[j]) is_a_min_Ix2_max_busy = true; + bool is_a_plus_Ix2_max_busy = false; + for (int j = 0; j < chain.Nstrings; ++j) if (plus_Ix2_max_busy[j]) is_a_plus_Ix2_max_busy = true; + */ + /* + // State is not admissible if all min_Ix2_max are busy simultaneously: + bool any_min_Ix2_max_free = false; + for (int j = 0; j < chain.Nstrings; ++j) + if (base[j] > 0 && !min_Ix2_max_busy[j]) any_min_Ix2_max_free = true; + if (!any_min_Ix2_max_free) return(false); + */ + /* + // State is not admissible if -Ix2_max, -Ix2_max + 2, ..., -Ix2_max + 2*(Str_L - 1) are busy: + for (int j = 0; j < chain.Nstrings; ++j) + if (base[j] > 0 && Ix2[j][0] <= -base.Ix2_max[j] + 2*(chain.Str_L[j] - 1)) + return(false); + // Almost correct with above ! + // State is not admissible if Ix2_max - 2, ..., Ix2_max - 2*(Str_L - 2) are busy (NB: one slot more than on left): + for (int j = 0; j < chain.Nstrings; ++j) + if (base[j] > 0 && Ix2[j][base[j] - 1] >= base.Ix2_max[j] - 2*(chain.Str_L[j] - 2)) + return(false); + */ + + // Check that at all at least doubly occupied levels, the difference between max and min quantum numbers + // is strictly smaller than 2*Ix2_max - 2, so that lambda_max - lambda_min < PI at each level: + //for (int j = 0; j < chain.Nstrings; ++j) + //if (base[j] >= 2 && Ix2[j][base[j] - 1] - Ix2[j][0] >= 2* base.Ix2_max[j] - 2) answer = false; + + //if (is_a_min_Ix2_max_busy && is_a_plus_Ix2_max_busy) return(false); + + bool higher_string_on_zero = false; + + for (int j = 0; j < chain.Nstrings; ++j) { + // The following line puts answer to true if there is at least one higher string with zero Ix2 + for (int alpha = 0; alpha < base[j]; ++alpha) if ((Ix2[j][alpha] == 0) && (chain.Str_L[j] > 2) /*&& !(chain.Str_L[j] % 2)*/) + higher_string_on_zero = true; + for (int alpha = 0; alpha < base[j]; ++alpha) if (Ix2[j][alpha] == 0) Zero_at_level[j] = true; + // NOTE: if base[j] == 0, Zero_at_level[j] remains false. + } + + bool symmetric_state = (*this).Check_Symmetry(); + + bool string_coincidence = false; + // Checks that we have strings of equal length modulo 2 and same parity with Ix2 == 0, so equal rapidities, and inadmissibility + for (int j1 = 0; j1 < chain.Nstrings; ++j1) { + for (int j2 = j1 + 1; j2 < chain.Nstrings; ++j2) + if (Zero_at_level[j1] && Zero_at_level[j2] && (!((chain.Str_L[j1] + chain.Str_L[j2])%2))) + string_coincidence = true; + } + + //bool onep_onem_on_zero = false; + //if (option == 'm' || option == 'p') { // for Smin, we also exclude symmetric states with 1+ and 1- strings on zero + //if (Zero_at_level[0] && Zero_at_level[1]) onep_onem_on_zero = true; + //} + + //cout << min_Ix2_max_busy << "\t" << symmetric_state << "\t" << higher_string_on_zero << "\t" << string_coincidence << "\t" << onep_onem_on_zero << endl; + + //answer = !((symmetric_state && (higher_string_on_zero || string_coincidence || onep_onem_on_zero))); + answer = !(symmetric_state && (higher_string_on_zero || string_coincidence)); + + // Explicitly exclude state with min_Ix2_max_busy && iK == 0 + //Compute_Momentum(); + //answer = !((min_Ix2_max_busy && iK == 0) || (symmetric_state && (higher_string_on_zero || string_coincidence || onep_onem_on_zero))); + + // Now check that no Ix2 is equal to +N (since we take -N into account, and I + N == I by periodicity of exp) + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) if ((Ix2[j][alpha] < -chain.Nsites) || (Ix2[j][alpha] >= chain.Nsites)) answer = false; + + if (!answer) { + E = 0.0; + K = 0.0; + conv = 0; + iter = 0; + iter_Newton = 0; + lnnorm = -100.0; + } + + return(answer); // answer == true: nothing wrong with this Ix2_config + } + + void XXZ_gpd_Bethe_State::Compute_BE (int j, int alpha) + { + // Fills in the BE members with the value of the Bethe equations. + + tanlambda[j][alpha] = tan(lambda[j][alpha]); + + DP sumtheta = 0.0; + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) { + sumtheta += atan ((tanlambda[j][alpha] - tanlambda[k][beta])/((1.0 + tanlambda[j][alpha] * tanlambda[k][beta]) + * chain.ta_n_anis_over_2[2])) + + PI * floor(0.5 + (lambda[j][alpha] - lambda[k][beta])/PI); + } + else sumtheta += 0.5 * Theta_XXZ_gpd((tanlambda[j][alpha] - tanlambda[k][beta])/(1.0 + tanlambda[j][alpha] * tanlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.ta_n_anis_over_2) + + PI * (2.0 * JSC::min(chain.Str_L[j], chain.Str_L[k]) - ((j == k) ? 1.0 : 0)) + * floor(0.5 + (lambda[j][alpha] - lambda[k][beta])/PI); + } + sumtheta *= 2.0; + + BE[j][alpha] = 2.0 * (atan(tanlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) + + PI * floor(0.5 + lambda[j][alpha]/PI)) + - (sumtheta + PI*Ix2[j][alpha])/chain.Nsites; + } + + void XXZ_gpd_Bethe_State::Compute_BE () + { + // Fills in the BE members with the value of the Bethe equations. + + (*this).Compute_tanlambda(); + + DP sumtheta = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) { + sumtheta += atan ((tanlambda[j][alpha] - tanlambda[k][beta])/((1.0 + tanlambda[j][alpha] * tanlambda[k][beta]) + * chain.ta_n_anis_over_2[2])) + + PI * floor(0.5 + (lambda[j][alpha] - lambda[k][beta])/PI); + } + else sumtheta += 0.5 * Theta_XXZ_gpd((tanlambda[j][alpha] - tanlambda[k][beta])/(1.0 + tanlambda[j][alpha] * tanlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.ta_n_anis_over_2) + + PI * (2.0 * JSC::min(chain.Str_L[j], chain.Str_L[k]) - ((j == k) ? 1.0 : 0)) + * floor(0.5 + (lambda[j][alpha] - lambda[k][beta])/PI); + } + sumtheta *= 2.0; + + BE[j][alpha] = 2.0 * (atan(tanlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) + + PI * floor(0.5 + lambda[j][alpha]/PI)) + - (sumtheta + PI*Ix2[j][alpha])/chain.Nsites; + } + } + + DP XXZ_gpd_Bethe_State::Iterate_BAE (int j, int alpha) + { + // Returns a new iteration value for lambda[j][alpha] given tanlambda[][] and BE[][] + // Assumes that tanlambda[][] and BE[][] have been computed. + + DP arg0 = 0.5 * (2.0 * (atan(tanlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) + + PI * floor(0.5 + lambda[j][alpha]/PI)) - BE[j][alpha]); + DP arg = chain.ta_n_anis_over_2[chain.Str_L[j]] * tan( + arg0 + //0.5 * + //(PI * Ix2[j][alpha] + sumtheta)/chain.Nsites + //(2.0 * (atan(tanlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) + // + PI * floor(0.5 + lambda[j][alpha]/PI)) - BE[j][alpha]) + ); + + return(atan(arg) + //+ PI * floor(0.5 + arg0) + //0.5 * (Ix2[j][alpha] + sumtheta/PI)/(chain.Nsites) + + PI * floor(0.5 + arg0/PI) + ); + + } + /* + void XXZ_gpd_Bethe_State::Iterate_BAE () + { + // Recalculates the rapidities by iterating Bethe equations + + Lambda New_lambda(chain, base); + DP sumtheta = 0.0; + DP arg = 0.0; + + // First, compute the tan of rapidities: + (*this).Compute_tanlambda(); + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += atan((tanlambda[j][alpha] - tanlambda[k][beta])/((1.0 + tanlambda[j][alpha] * tanlambda[k][beta]) + * chain.ta_n_anis_over_2[2])) + + PI * floor(0.5 + (lambda[j][alpha] - lambda[k][beta])/PI); + + else sumtheta += 0.5 * Theta_XXZ_gpd((tanlambda[j][alpha] - tanlambda[k][beta])/(1.0 + tanlambda[j][alpha] * tanlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.ta_n_anis_over_2) + + PI * (2.0 * JSC::min(chain.Str_L[j], chain.Str_L[k]) - ((j == k) ? 1.0 : 0)) + * floor(0.5 + (lambda[j][alpha] - lambda[k][beta])/PI); + } + sumtheta *= 2.0; + + arg = chain.ta_n_anis_over_2[chain.Str_L[j]] * tan((PI * 0.5 * Ix2[j][alpha] + 0.5 * sumtheta)/chain.Nsites); + + New_lambda[j][alpha] = atan(arg) + PI * floor(0.5 + (0.5 * Ix2[j][alpha] + 0.5 * sumtheta/PI)/(chain.Nsites)); + + } + } + + DP New_diffsq = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + //New_diffsq += pow(tan(New_lambda[j][alpha]) - tanlambda[j][alpha], 2.0); + New_diffsq += pow(New_lambda[j][alpha] - lambda[j][alpha], 2.0); + lambda[j][alpha] = 1.0 * New_lambda[j][alpha] + 0.0 * lambda[j][alpha]; + } + } + + diffsq = New_diffsq; + iter++; + + return; + } + */ + /* + void XXZ_gpd_Bethe_State::Iterate_BAE_Newton () + { + // does one step of a Newton method on the rapidities... + + Vect_DP RHSBAE (0.0, base.Nraptot); // contains RHS of BAEs + Vect_CX dlambda (0.0, base.Nraptot); // contains delta lambda computed from Newton's method + SQMat_CX Gaudin (0.0, base.Nraptot); + Vect_INT indx (base.Nraptot); + DP sumtheta = 0.0; + DP arg = 0.0; + DP fn_arg = 0.0; + DP olddiffsq = diffsq; + + // Compute the RHS of the BAEs: + + int index = 0; + + (*this).Compute_tanlambda(); + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) { + sumtheta += atan ((tanlambda[j][alpha] - tanlambda[k][beta])/((1.0 + tanlambda[j][alpha] * tanlambda[k][beta]) + * chain.ta_n_anis_over_2[2])) + + PI * floor(0.5 + (lambda[j][alpha] - lambda[k][beta])/PI); + } + else sumtheta += 0.5 * Theta_XXZ_gpd((tanlambda[j][alpha] - tanlambda[k][beta])/(1.0 + tanlambda[j][alpha] * tanlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.ta_n_anis_over_2) + + PI * (2.0 * JSC::min(chain.Str_L[j], chain.Str_L[k]) - ((j == k) ? 1.0 : 0)) + * floor(0.5 + (lambda[j][alpha] - lambda[k][beta])/PI); + } + sumtheta *= 2.0; + + RHSBAE[index] = chain.Nsites * 2.0 * (atan(tanlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) + + PI * floor(0.5 + lambda[j][alpha]/PI)) + // ) + - sumtheta - PI*Ix2[j][alpha]; + index++; + } + } + + (*this).Build_Reduced_Gaudin_Matrix (Gaudin); + + for (int i = 0; i < base.Nraptot; ++i) dlambda[i] = - RHSBAE[i]; + + DP d; + ludcmp_CX (Gaudin, indx, d); + lubksb_CX (Gaudin, indx, dlambda); + + diffsq = 0.0; + // for (int i = 0; i < base.Nraptot; ++i) diffsq += norm(dlambda[i]); + int ctr = 0; + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + diffsq += norm(tan(lambda[j][alpha] + dlambda[ctr]) - tanlambda[j][alpha]); + // cout << "lambda = " << lambda[j][alpha] << "\tdlambda = " << dlambda[ctr] << endl; + ctr++; + } + } + + // if we've converged, calculate the norm here, since the work has been done... + + if (diffsq < chain.prec) { + lnnorm = 0.0; + for (int j = 0; j < base.Nraptot; j++) lnnorm += log(abs(Gaudin[j][j])); + } + + index = 0; + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + lambda[j][alpha] += real(dlambda[index]); + index++; + } + } + + iter_Newton++; + + // cout << "iter_N = " << iter_Newton << "\t" << diffsq << endl; + + return; + } + */ + bool XXZ_gpd_Bethe_State::Check_Rapidities() + { + bool nonan = true; + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) if (nonan) nonan = ((!is_nan(lambda[j][alpha])) + //&& (lambda[j][alpha] > -0.5*PI*chain.Str_L[j]) + //&& (lambda[j][alpha] <= 0.5*PI*chain.Str_L[j]) + ); + + bool all_within_pi_interval = true; + DP min_lambda = 10.0; + DP max_lambda = -10.0; + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + if (lambda[j][alpha] < min_lambda) min_lambda = lambda[j][alpha]; + if (lambda[j][alpha] > max_lambda) max_lambda = lambda[j][alpha]; + } + + //all_within_pi_interval = (fabs(max_lambda - min_lambda) < PI - 0.0 * 1.0e-12); + // Use pi interval, but centered on 0: + all_within_pi_interval = max_lambda < 0.5* PI + 1.0e-8 && min_lambda > -0.5* PI - 1.0e-8; + //nonan *= all_within_pi_interval; + + return nonan; + } + + DP XXZ_gpd_Bethe_State::String_delta () + { + // Computes the sum of absolute value of \delta^{a, a+1} in string hypothesis, for a given bethe eigenstate + + DP delta = 0.0; + + int occupied_strings = 0; + for (int i = 0; i < (*this).chain.Nstrings; ++i) if ((*this).chain.Str_L[i] > 1) occupied_strings += (*this).base.Nrap[i]; + + //if ((*this).conv == 0) delta = 1.0; + + if (occupied_strings == 0) delta = 0.0; + + else { + + Vect_DP ln_deltadiff(0.0, 1000); // contains ln |delta^{a, a+1}| + Vect_DP deltadiff(0.0, 1000); // contains |delta^{a, a+1}| + + complex log_BAE_reg = 0.0; + + for (int j = 0; j < (*this).chain.Nstrings; ++j) { + for (int alpha = 0; alpha < (*this).base[j]; ++alpha) { + + ln_deltadiff = 0.0; + + for (int a = 1; a <= (*this).chain.Str_L[j]; ++a) { + + if ((*this).chain.Str_L[j] > 1) { // else the BAE are already 1 + + log_BAE_reg = DP((*this).chain.Nsites) * log(sin((*this).lambda[j][alpha] + 0.5 * II * (*this).chain.anis + * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a + 1.0)) + /sin((*this).lambda[j][alpha] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a - 1.0))); + + for (int k = 0; k < (*this).chain.Nstrings; ++k) + for (int beta = 0; beta < (*this).base[k]; ++beta) + for (int b = 1; b <= (*this).chain.Str_L[k]; ++b) { + if ((j != k) || (alpha != beta) || (a != b - 1)) + + log_BAE_reg += log(sin(((*this).lambda[j][alpha] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a )) + - ((*this).lambda[k][beta] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[k] + 1.0 - 2.0 * b )) + - II * (*this).chain.anis)); + + if ((j != k) || (alpha != beta) || (a != b + 1)) + + log_BAE_reg -= log(sin(((*this).lambda[j][alpha] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[j] + 1.0 - 2.0 * a )) + - ((*this).lambda[k][beta] + 0.5 * II * (*this).chain.anis * ((*this).chain.Str_L[k] + 1.0 - 2.0 * b )) + + II * (*this).chain.anis)); + } + + // The regular LHS of BAE is now defined. Now sum up the deltas... + + if (a == 1) ln_deltadiff[0] = -real(log_BAE_reg); + + else if (a < (*this).chain.Str_L[j]) ln_deltadiff[a - 1] = ln_deltadiff[a-2] - real(log_BAE_reg); + + else if (a == (*this).chain.Str_L[j]) ln_deltadiff[a-1] = real(log_BAE_reg); + + } // if ((*this).chain.Str_L[j] > 1) + + } // for (int a = 1; ... + + for (int a = 0; a < (*this).chain.Str_L[j]; ++a) { + deltadiff[a] = ln_deltadiff[a] != 0.0 ? exp(ln_deltadiff[a]) : 0.0; + delta += fabs(deltadiff[a]); + } + + } // alpha sum + } // j sum + + if (is_nan(delta)) delta = 1.0; // sentinel + + } // else + + return delta; + } + + + void XXZ_gpd_Bethe_State::Compute_Energy () + { + DP sum = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum += sinh(chain.Str_L[j] * chain.anis) / (cos(2.0 * lambda[j][alpha]) - cosh(chain.Str_L[j] * chain.anis)); + } + } + + sum *= chain.J * sinh(chain.anis); + + E = sum; + + return; + } + + /* + void XXZ_gpd_Bethe_State::Compute_Momentum () + { + int sum_Ix2 = 0; + DP sum_M = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + sum_M += base[j]; + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum_Ix2 += Ix2[j][alpha]; + } + } + + iK = (chain.Nsites/2) * int(sum_M) - (sum_Ix2/2); + + while (iK >= chain.Nsites) iK -= chain.Nsites; + while (iK < 0) iK += chain.Nsites; + + K = PI * sum_M - PI * sum_Ix2/chain.Nsites; + + while (K >= 2.0*PI) K -= 2.0*PI; + while (K < 0.0) K += 2.0*PI; + + return; + } + */ + + void XXZ_gpd_Bethe_State::Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red) + { + + if (Gaudin_Red.size() != base.Nraptot) JSCerror("Passing matrix of wrong size in Build_Reduced_Gaudin_Matrix."); + + int index_jalpha; + int index_kbeta; + + DP sum_hbar_XXZ = 0.0; + + DP sinhetasq = pow(sinh(chain.anis), 2.0); + + (*this).Compute_sinlambda(); + (*this).Compute_coslambda(); + + index_jalpha = 0; + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + index_kbeta = 0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) { + + if ((j == k) && (alpha == beta)) { + + sum_hbar_XXZ = 0.0; + + for (int kp = 0; kp < chain.Nstrings; ++kp) { + for (int betap = 0; betap < base[kp]; ++betap) { + if (!((j == kp) && (alpha == betap))) + sum_hbar_XXZ + += ddlambda_Theta_XXZ_gpd (lambda[j][alpha] - lambda[kp][betap], chain.Str_L[j], chain.Str_L[kp], + chain.si_n_anis_over_2); + } + } + + Gaudin_Red[index_jalpha][index_kbeta] + = complex ( chain.Nsites * hbar_XXZ_gpd (lambda[j][alpha], chain.Str_L[j], chain.si_n_anis_over_2) - sum_hbar_XXZ); + } + + else { + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + Gaudin_Red[index_jalpha][index_kbeta] = + complex (chain.si_n_anis_over_2[4]/(pow(sinlambda[j][alpha] * coslambda[k][beta] + - coslambda[j][alpha] * sinlambda[k][beta], 2.0) + sinhetasq)); + else + Gaudin_Red[index_jalpha][index_kbeta] = complex (ddlambda_Theta_XXZ_gpd (lambda[j][alpha] - lambda[k][beta], chain.Str_L[j], + chain.Str_L[k], chain.si_n_anis_over_2)); + } + index_kbeta++; + } + } + index_jalpha++; + } + } + + return; + } + + + // **************************************************************************************************** + + // non-member functions + + inline DP fbar_XXZ_gpd (DP tanlambda, DP tanhnetaover2) + { + return (2.0 * atan(tanlambda/tanhnetaover2)); + } + + DP Theta_XXZ_gpd (DP tanlambda, int nj, int nk, DP* tanhnetaover2) + { + DP result; + + if ((nj == 1) && (nk == 1)) result = fbar_XXZ_gpd(tanlambda, tanhnetaover2[2]); + + else { + + result = (nj == nk) ? 0.0 : fbar_XXZ_gpd(tanlambda, tanhnetaover2[fabs(nj - nk)]); + + for (int a = 1; a < JSC::min(nj, nk); ++a) result += 2.0 * fbar_XXZ_gpd(tanlambda, tanhnetaover2[fabs(nj - nk) + 2*a]); + + result += fbar_XXZ_gpd(tanlambda, tanhnetaover2[nj + nk]); + } + + return (result); + } + + DP hbar_XXZ_gpd (DP lambda, int n, DP* si_n_anis_over_2) + { + return (si_n_anis_over_2[2*n]/(pow(sin(lambda), 2.0) + pow(si_n_anis_over_2[n], 2.0))); + } + + DP ddlambda_Theta_XXZ_gpd (DP lambda, int nj, int nk, DP* si_n_anis_over_2) + { + DP result = (nj == nk) ? 0.0 : hbar_XXZ_gpd(lambda, fabs(nj - nk), si_n_anis_over_2); + + for (int a = 1; a < JSC::min(nj, nk); ++a) result += 2.0 * hbar_XXZ_gpd(lambda, fabs(nj - nk) + 2*a, si_n_anis_over_2); + + result += hbar_XXZ_gpd(lambda, nj + nk, si_n_anis_over_2); + + return (result); + } + + + XXZ_gpd_Bethe_State Add_Particle_at_Center (const XXZ_gpd_Bethe_State& RefState) + { + if (2*RefState.base.Mdown == RefState.chain.Nsites) + JSCerror("Trying to add a down spin to a zero-magnetized chain in Add_Particle_at_Center."); + + Vect newM = RefState.base.Nrap; + newM[0] = newM[0] + 1; + + Heis_Base newBase (RefState.chain, newM); + + XXZ_gpd_Bethe_State ReturnState (RefState.chain, newBase); + + for (int il = 1; il < RefState.chain.Nstrings; ++il) + for (int alpha = 0; alpha < RefState.base.Nrap[il]; ++alpha) + ReturnState.Ix2[il][alpha] = RefState.Ix2[il][alpha]; + + // Add a quantum number in middle (explicitly: to right of index M[0]/2) + // and shift quantum numbers by half-integer away from added one: + ReturnState.Ix2[0][RefState.base.Nrap[0]/2] = RefState.Ix2[0][RefState.base.Nrap[0]/2] - 1; + for (int i = 0; i < RefState.base.Nrap[0] + 1; ++i) + ReturnState.Ix2[0][i + (i >= RefState.base.Nrap[0]/2)] = RefState.Ix2[0][i] - 1 + 2*(i >= RefState.base.Nrap[0]/2); + + return(ReturnState); + } + + + XXZ_gpd_Bethe_State Remove_Particle_at_Center (const XXZ_gpd_Bethe_State& RefState) + { + if (RefState.base.Nrap[0] == 0) + JSCerror("Trying to remove a down spin in an empty Nrap[0] state."); + + Vect newM = RefState.base.Nrap; + newM[0] = newM[0] - 1; + + Heis_Base newBase (RefState.chain, newM); + + XXZ_gpd_Bethe_State ReturnState (RefState.chain, newBase); + + for (int il = 1; il < RefState.chain.Nstrings; ++il) + for (int alpha = 0; alpha < RefState.base.Nrap[il]; ++alpha) + ReturnState.Ix2[il][alpha] = RefState.Ix2[il][alpha]; + + // Remove midmost and shift quantum numbers by half-integer towards removed one: + for (int i = 0; i < RefState.base.Nrap[0]-1; ++i) + ReturnState.Ix2[0][i] = RefState.Ix2[0][i + (i >= RefState.base.Nrap[0]/2)] + 1 - 2*(i >= RefState.base.Nrap[0]/2); + + return(ReturnState); + } + +} // namespace JSC diff --git a/src/HEIS/ln_Overlap_XXX.cc b/src/HEIS/ln_Overlap_XXX.cc new file mode 100644 index 0000000..7569aba --- /dev/null +++ b/src/HEIS/ln_Overlap_XXX.cc @@ -0,0 +1,253 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +***********************************************************/ + +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + +inline complex ln_Fn_F (XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) + ans += log(B.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex ln_Fn_G (XXX_Bethe_State& A, XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + ans += log(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex Fn_K (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return(1.0/((A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0)) ))); +} + +inline complex Fn_L (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return ((2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + )) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + +complex ln_Overlap (XXX_Bethe_State& A, XXX_Bethe_State& B) +{ + // This function returns the overlap of states A and B. + // The A and B states can contain strings. + + // IMPORTANT ASSUMPTIONS: + // - State B is an eigenstate of the model on which the overlap measure is defined + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown) return(complex(-300.0)); // overlap vanishes + + // Some convenient arrays + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + //complex ln_prod1 = 0.0; + //complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + /* + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm((A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm((B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm((B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)))); + */ + + // Define the F ones earlier... + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) + ln_prod3 += ln_Fn_F (A, j, alpha, a - 1); + + // ln_prod3 -= A.base.Mdown * log(abs(sin(A.chain.zeta))); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + // ln_prod4 -= B.base.Mdown * log(abs(sin(B.chain.zeta))); + + // Now proceed to build the Hm2P matrix + + SQMat_CX Hm2P(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + //complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + //complex two_over_A_lambda_sq_plus_1over2sq; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + //two_over_A_lambda_sq_plus_1over2sq = 2.0/((A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) * + // (A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) + 0.25); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm2P matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta]); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta]); + + //Prod_powerN = pow((B.lambda[k][beta] + 0.5 * II)/(B.lambda[k][beta] - 0.5 * II), complex (B.chain.Nsites)); + Prod_powerN = pow((B.lambda[k][beta] + 0.5 * II)/(B.lambda[k][beta] - 0.5 * II), complex (A.chain.Nsites)); // careful ! + + Hm2P[index_a][index_b] = Fn_K_0_G_0 - Prod_powerN * Fn_K_1_G_2 + //- two_over_A_lambda_sq_plus_1over2sq * exp(II*im_ln_Fn_F_B_0[k][beta]); + ; + } + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm2P[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + //sum2 = 0.0; + + //for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]]); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_Fn_F(B, k, beta, jsum - 1))); // include all string contributions F_B_0 in this term + + //Hm2P[index_a][index_b] = prod_num * (sum1 - sum2 * two_over_A_lambda_sq_plus_1over2sq); + Hm2P[index_a][index_b] = prod_num * sum1; + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + + index_a++; + }}} // sums over j, alpha, a + + //cout << "Matrix: " << endl; + //Hm2P.Print(); + + complex det = lndet_LU_CX_dstry(Hm2P); + + /* + complex ln_form_factor_sq = log(0.25 * A.chain.Nsites) + real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) + // + 2.0 * real(lndet_LU_CX_dstry(Hm2P)) + + 2.0 * det + - A.lnnorm - B.lnnorm; + + //cout << "ln_SZ: " << endl << ln_prod1 << "\t" << -ln_prod2 << "\t" << -ln_prod3 << "\t" << ln_prod4 << "\t" << 2.0 * det + // << "\t" << -A.lnnorm << "\t" << -B.lnnorm << endl; + + return(ln_form_factor_sq); + */ + complex ln_overlap = 0.5 * (-ln_prod3 + ln_prod4) + det - 0.5 * (A.lnnorm + B.lnnorm); + + cout << "ln_overlap: " << endl << -ln_prod3 << "\t" << ln_prod4 << "\t" << 2.0 * det + << "\t" << -A.lnnorm << "\t" << -B.lnnorm << endl; + + return(ln_overlap); + +} + +} // namespace JSC diff --git a/src/HEIS/ln_Smin_ME_XXX.cc b/src/HEIS/ln_Smin_ME_XXX.cc new file mode 100644 index 0000000..5162573 --- /dev/null +++ b/src/HEIS/ln_Smin_ME_XXX.cc @@ -0,0 +1,239 @@ +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + +inline complex ln_Fn_F (XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) + ans += log(B.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex ln_Fn_G (XXX_Bethe_State& A, XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + ans += log(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex Fn_K (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return(1.0/((A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0)) ))); +} + +inline complex Fn_L (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return ((2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + )) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + +complex ln_Smin_ME (XXX_Bethe_State& A, XXX_Bethe_State& B) +{ + // This function returns the natural log of the S^- operator matrix element. + // The A and B states can contain strings. + + // Check that the two states are compatible + + if (A.chain != B.chain) JSCerror("Incompatible XXX_Chains in Smin matrix element."); + + // Check that A and B are Mdown-compatible: + + if (A.base.Mdown != B.base.Mdown + 1) JSCerror("Incompatible Mdown between the two states in Smin matrix element!"); + + // Some convenient arrays + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))); + + // Define the F ones earlier... + + complex ln_FB0, ln_FG0, ln_FG2; + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + //re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + //im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + ln_FB0 = ln_Fn_F(B, j, alpha, 0); + re_ln_Fn_F_B_0[j][alpha] = real(ln_FB0); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_FB0); + //re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + //im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + ln_FG0 = ln_Fn_G(A, B, j, alpha, 0); + re_ln_Fn_G_0[j][alpha] = real(ln_FG0); + im_ln_Fn_G_0[j][alpha] = imag(ln_FG0); + //re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + //im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + ln_FG2 = ln_Fn_G(A, B, j, alpha, 2); + re_ln_Fn_G_2[j][alpha] = real(ln_FG2); + im_ln_Fn_G_2[j][alpha] = imag(ln_FG2); + } + } + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) + ln_prod3 += ln_Fn_F(A, j, alpha, a - 1); + + // ln_prod3 -= A.base.Mdown * log(abs(sin(A.chain.zeta))); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + // ln_prod4 -= B.base.Mdown * log(abs(sin(B.chain.zeta))); + + // Now proceed to build the Hm matrix + + SQMat_CX Hm(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + complex one_over_A_lambda_sq_plus_1over2sq; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + one_over_A_lambda_sq_plus_1over2sq = 1.0/((A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) * + (A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) + 0.25); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm2P matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta]); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta]); + + Prod_powerN = pow((B.lambda[k][beta] + II * 0.5) /(B.lambda[k][beta] - II * 0.5), complex (B.chain.Nsites)); + + Hm[index_a][index_b] = Fn_K_0_G_0 - Prod_powerN * Fn_K_1_G_2; + + } // if (B.chain.Str_L == 1) + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + /* + sum2 = 0.0; + + for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + + */ + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]]); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_Fn_F(B, k, beta, jsum - 1))); + // include all string contributions F_B_0 in this term + + Hm[index_a][index_b] = prod_num * sum1; + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + + // now define the elements Hm[a][M] + + Hm[index_a][B.base.Mdown] = one_over_A_lambda_sq_plus_1over2sq; + + index_a++; + }}} // sums over j, alpha, a + + complex ln_ME_sq = log(1.0 * A.chain.Nsites) + real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) + + 2.0 * real(lndet_LU_CX_dstry(Hm)) - A.lnnorm - B.lnnorm; + + //return(ln_ME_sq); + return(0.5 * ln_ME_sq); // Return ME, not MEsq + +} + +} // namespace JSC diff --git a/src/HEIS/ln_Smin_ME_XXZ.cc b/src/HEIS/ln_Smin_ME_XXZ.cc new file mode 100644 index 0000000..6a3aef7 --- /dev/null +++ b/src/HEIS/ln_Smin_ME_XXZ.cc @@ -0,0 +1,328 @@ +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + +inline complex ln_Fn_F (XXZ_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + int arg = 0; + int absarg = 0; + int par_comb_1, par_comb_2; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + + par_comb_1 = B.chain.par[j] == B.chain.par[k] ? 1 : 0; + par_comb_2 = B.chain.par[k] == B.chain.par[j] ? 0 : B.chain.par[k]; + + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) { + + arg = B.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + absarg = abs(arg); + /* + prod_temp *= 0.5 * //done later... + ((B.sinhlambda[j][alpha] * B.coshlambda[k][beta] - B.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (B.chain.co_n_anis_over_2[absarg] * (1.0 + B.chain.par[j] * B.chain.par[k]) + - sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * (B.chain.par[k] - B.chain.par[j])) + + II * (B.coshlambda[j][alpha] * B.coshlambda[k][beta] - B.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * (1.0 + B.chain.par[j] * B.chain.par[k]) + + B.chain.co_n_anis_over_2[absarg] * (B.chain.par[k] - B.chain.par[j])) ); + */ + + prod_temp *= ((B.sinhlambda[j][alpha] * B.coshlambda[k][beta] - B.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (B.chain.co_n_anis_over_2[absarg] * par_comb_1 - sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * par_comb_2) + + II * (B.coshlambda[j][alpha] * B.coshlambda[k][beta] - B.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * par_comb_1 + B.chain.co_n_anis_over_2[absarg] * par_comb_2)); + } + + if (counter++ > 100) { // we do at most 100 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + + }}} + + return(ans + log(prod_temp)); +} + +inline complex ln_Fn_G (XXZ_Bethe_State& A, XXZ_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + int arg = 0; + int absarg = 0; + int par_comb_1, par_comb_2; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + + par_comb_1 = A.chain.par[j] == B.chain.par[k] ? 1 : 0; + par_comb_2 = B.chain.par[k] == A.chain.par[j] ? 0 : B.chain.par[k]; + + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + arg = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + absarg = abs(arg); + /* + prod_temp *= 0.5 * //done later... + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg] * (B.chain.par[k] - A.chain.par[j])) ); + */ + prod_temp *= ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg] * par_comb_1 - sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * par_comb_2) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * par_comb_1 + A.chain.co_n_anis_over_2[absarg] * par_comb_2)); + + if (counter++ > 100) { // we do at most 100 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + }}} + + return(ans + log(prod_temp)); +} + +inline complex Fn_K (XXZ_Bethe_State& A, int j, int alpha, int a, XXZ_Bethe_State& B, int k, int beta, int b) +{ + int arg1 = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + int absarg1 = abs(arg1); + int arg2 = arg1 + 2; + int absarg2 = abs(arg2); + + return(4.0/( + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg1] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg1) * A.chain.si_n_anis_over_2[absarg1] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg1) * A.chain.si_n_anis_over_2[absarg1] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg1] * (B.chain.par[k] - A.chain.par[j])) ) + * + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg2] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg2) * A.chain.si_n_anis_over_2[absarg2] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg2) * A.chain.si_n_anis_over_2[absarg2] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg2] * (B.chain.par[k] - A.chain.par[j])) ) + )); + +} + +inline complex Fn_L (XXZ_Bethe_State& A, int j, int alpha, int a, XXZ_Bethe_State& B, int k, int beta, int b) +{ + return (sinh(2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + + 0.25 * II * PI * complex(-A.chain.par[j] + B.chain.par[k]))) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + +complex ln_Smin_ME (XXZ_Bethe_State& A, XXZ_Bethe_State& B) +{ + // This function returns the natural log of the S^- operator matrix element. + // The A and B states can contain strings. + + // Check that the two states are compatible + + if (A.chain != B.chain) JSCerror("Incompatible XXZ_Chains in Smin matrix element."); + + // Check that A and B are Mdown-compatible: + + if (A.base.Mdown != B.base.Mdown + 1) { + cout << "A.base.Mdown = " << A.base.Mdown << "\tB.base.Mdown = " << B.base.Mdown << endl; + cout << "A: " << A << endl << "B: " << B << endl; + JSCerror("Incompatible Mdown between the two states in Smin matrix element!"); + } + + // Compute the sinh and cosh of rapidities + + A.Compute_sinhlambda(); + A.Compute_coshlambda(); + B.Compute_sinhlambda(); + B.Compute_coshlambda(); + + // Some convenient arrays + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(sinh(A.lambda[i][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - A.chain.par[i])))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - B.chain.par[i]))) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - B.chain.par[i])))); + + // Define the F ones earlier... + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + DP logabssinzeta = log(abs(sin(A.chain.anis))); + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) + ln_prod3 += ln_Fn_F(A, j, alpha, a - 1); // assume only one-strings here + + ln_prod3 -= A.base.Mdown * log(abs(sin(A.chain.anis))); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + ln_prod4 -= B.base.Mdown * log(abs(sin(B.chain.anis))); + + // Now proceed to build the Hm matrix + + SQMat_CX Hm(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + complex one_over_A_sinhlambda_sq_plus_sinzetaover2sq; + + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + one_over_A_sinhlambda_sq_plus_sinzetaover2sq = 1.0/((sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + * (sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + + pow(sin(0.5*A.chain.anis), 2.0)); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinzeta); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinzeta); + + Prod_powerN = pow( B.chain.par[k] == 1 ? + (B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + : + (B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + , complex (B.chain.Nsites)); + + Hm[index_a][index_b] = Fn_K_0_G_0 - Prod_powerN * Fn_K_1_G_2; + + } // if (B.chain.Str_L == 1) + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + /* + sum2 = 0.0; + + for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + + */ + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]] + logabssinzeta); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_Fn_F(B, k, beta, jsum - 1)) + logabssinzeta); + // include all string contributions F_B_0 in this term + + Hm[index_a][index_b] = prod_num * sum1; + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + + // now define the elements Hm[a][M] + + Hm[index_a][B.base.Mdown] = one_over_A_sinhlambda_sq_plus_sinzetaover2sq; + + index_a++; + }}} // sums over j, alpha, a + + complex ln_ME_sq = log(1.0 * A.chain.Nsites) + real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) + + 2.0 * real(lndet_LU_CX_dstry(Hm)) + logabssinzeta - A.lnnorm - B.lnnorm; + + //return(ln_ME_sq); + return(0.5 * ln_ME_sq); // Return ME, not MEsq + +} + +} // namespace JSC diff --git a/src/HEIS/ln_Smin_ME_XXZ_gpd.cc b/src/HEIS/ln_Smin_ME_XXZ_gpd.cc new file mode 100644 index 0000000..f5c93be --- /dev/null +++ b/src/HEIS/ln_Smin_ME_XXZ_gpd.cc @@ -0,0 +1,303 @@ +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + +inline complex ln_Fn_F (XXZ_gpd_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + /* + if (!((j == k) && (alpha == beta) && (a == b))) + ans += log(-II * sin(B.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.eta * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b)))); + */ + + if (!((j == k) && (alpha == beta) && (a == b))) { + + // arg = B.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + // absarg = abs(arg); + + prod_temp *= -II * (B.sinlambda[j][alpha] * B.coslambda[k][beta] - B.coslambda[j][alpha] * B.sinlambda[k][beta]) + * cosh(0.5 * B.chain.anis * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + + (B.coslambda[j][alpha] * B.coslambda[k][beta] + B.sinlambda[j][alpha] * B.sinlambda[k][beta]) + * sinh(0.5 * B.chain.anis * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + + } + + + if (counter++ > 10) { // we do at most 10 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + + }}} + + // return(ans); + return(ans + log(prod_temp)); +} + +inline complex ln_Fn_G (XXZ_gpd_Bethe_State& A, XXZ_gpd_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + /* + prod_temp *= -II * sin(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.eta * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + */ + + // arg = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + // absarg = abs(arg); + + prod_temp *= -II * (A.sinlambda[j][alpha] * B.coslambda[k][beta] - A.coslambda[j][alpha] * B.sinlambda[k][beta]) + * cosh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + + (A.coslambda[j][alpha] * B.coslambda[k][beta] + A.sinlambda[j][alpha] * B.sinlambda[k][beta]) + * sinh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + + if (counter++ > 25) { // we do at most 25 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + }}} + + // return(ans); + return(ans + log(prod_temp)); +} + +inline complex Fn_K (XXZ_gpd_Bethe_State& A, int j, int alpha, int a, XXZ_gpd_Bethe_State& B, int k, int beta, int b) +{ + /* + return(-1.0/(sin(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.zeta * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + * sin(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.zeta * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0))))); + */ + + // int arg1 = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + // int absarg1 = abs(arg1); + // int arg2 = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b - 1); + // int arg2 = arg1 + 2; + // int absarg2 = abs(arg2); + + return(1.0/(( -II * (A.sinlambda[j][alpha] * B.coslambda[k][beta] - A.coslambda[j][alpha] * B.sinlambda[k][beta]) + * cosh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + + (A.coslambda[j][alpha] * B.coslambda[k][beta] + A.sinlambda[j][alpha] * B.sinlambda[k][beta]) + * sinh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b)))) * + (-II * (A.sinlambda[j][alpha] * B.coslambda[k][beta] - A.coslambda[j][alpha] * B.sinlambda[k][beta]) + * cosh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0))) + + (A.coslambda[j][alpha] * B.coslambda[k][beta] + A.sinlambda[j][alpha] * B.sinlambda[k][beta]) + * sinh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0)))))); + +} + +inline complex Fn_L (XXZ_gpd_Bethe_State& A, int j, int alpha, int a, XXZ_gpd_Bethe_State& B, int k, int beta, int b) +{ + /* + complex ans = -II * sin(2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.zeta * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)))); + + return (ans * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); + */ + return (-II * sin(2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)))) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + +complex ln_Smin_ME (XXZ_gpd_Bethe_State& A, XXZ_gpd_Bethe_State& B) +{ + // This function returns the natural log of the S^- operator matrix element. + // The A and B states can contain strings. + + // Check that the two states refer to the same XXZ_gpd_Chain + + if (A.chain != B.chain) JSCerror("Incompatible XXZ_gpd_Chains in Smin matrix element."); + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown + 1) JSCerror("Incompatible Mdown between the two states in Smin matrix element!"); + + // Compute the sin and cos of rapidities + + A.Compute_sinlambda(); + A.Compute_coslambda(); + B.Compute_sinlambda(); + B.Compute_coslambda(); + + // Some convenient arrays + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(sin(A.lambda[i][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(sin(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(sin(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)))); + + // one_over_A_sinlambda_sq_plus_sinhetaover2sq[j][alpha] = 1.0/(pow(A.sinlambda[j][alpha], 2.0) + pow(sinh(0.5*A.chain.anis), 2.0)); + + // Define the F ones earlier... + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + DP logabssinheta = log(abs(sinh(A.chain.anis))); + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) + ln_prod3 += ln_Fn_F(A, j, alpha, a - 1); + + ln_prod3 -= A.base.Mdown * log(abs(sinh(A.chain.anis))); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + ln_prod4 -= B.base.Mdown * log(abs(sinh(B.chain.anis))); + + // Now proceed to build the Hm matrix + + SQMat_CX Hm(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + complex one_over_A_sinlambda_sq_plus_sinhetaover2sq = 0.0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + one_over_A_sinlambda_sq_plus_sinhetaover2sq = -1.0/((sin(A.lambda[j][alpha] // minus sign: from 1/(sinh^2... to -1/(sin^2... + + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a))) * + (sin(A.lambda[j][alpha] // minus sign: from 1/(sinh^2... to -1/(sin^2... + + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a))) + + pow(sinh(0.5*A.chain.anis), 2.0)); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm2P matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinheta); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinheta); + + Prod_powerN = pow((B.sinlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.coslambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.sinlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.coslambda[k][beta] * B.chain.si_n_anis_over_2[1]), + complex (B.chain.Nsites)); + + Hm[index_a][index_b] = Fn_K_0_G_0 - Prod_powerN * Fn_K_1_G_2; + } + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + /* + sum2 = 0.0; + + for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + */ + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]] + logabssinheta); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_Fn_F(B, k, beta, jsum - 1)) + logabssinheta); + // include all string contributions F_B_0 in this term + + Hm[index_a][index_b] = prod_num * sum1; + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + + // now define the elements Hm[a][M] + + Hm[index_a][B.base.Mdown] = one_over_A_sinlambda_sq_plus_sinhetaover2sq; + + index_a++; + }}} // sums over j, alpha, a + + complex ln_ME_sq = log(1.0 * A.chain.Nsites) + real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) + + 2.0 * real(lndet_LU_CX_dstry(Hm)) + logabssinheta - A.lnnorm - B.lnnorm; + + //return(ln_ME_sq); + return(0.5 * ln_ME_sq); // Return ME, not MEsq + +} + +} // namespace JSC diff --git a/src/HEIS/ln_Smm_ME_XXX.cc b/src/HEIS/ln_Smm_ME_XXX.cc new file mode 100644 index 0000000..7975319 --- /dev/null +++ b/src/HEIS/ln_Smm_ME_XXX.cc @@ -0,0 +1,307 @@ +#include "JSC.h" + +using namespace JSC; + +namespace JSC { +inline complex phi(complex x){return x;} +inline complex a(complex x){return 1;} +inline complex b(complex x,complex y, complex eta){ return phi(x-y)/phi(x-y+complex(0.0,1.0)*eta);} +inline complex d(complex x, complex xi, complex eta, int N){return pow(b(x,xi,eta),N);} + + +inline complex ln_Fn_F (XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) + ans += log(B.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex ln_Fn_G (XXX_Bethe_State& A, XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + ans += log(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + + } + } + + } + + return(ans); +} + +inline complex Fn_K (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + + return(1.0/((A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0)) ))); + +} + +inline complex Fn_L (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return ((2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + )) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + + +complex ln_Smm_ME (XXX_Bethe_State& A, XXX_Bethe_State& B) +{ + const DP Zero_Center_Thres=1.0e-5; +const DP real_dev=1.0e-14; + const DP Diff_ME_Thres=1.e-6; + + //clock_t start_time_local = clock(); + + // This function returns the natural log of the S^z operator matrix element. + // The A and B states can contain strings. +// A is the reference state. + + // Check that the two states refer to the same XXX_Chain + + if (A.chain != B.chain) JSCerror("Incompatible XXX_Chains in Smm matrix element."); + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown + 2) JSCerror("Incompatible Mdown between the two states in Smm matrix element!"); + + //if(B.String_delta()> HEIS_deltaprec) return(complex(-300.0)); // DEPRECATED in ++T_9 + + + //if (B.type_id > 999999LL) return(complex(-300.0)); + + // Some convenient arrays + complex eta=-II; + complex ln_prod = complex(0.0,0.0); + complex result=-300; + complex prev_result=-300; + +XXX_Bethe_State B_origin; B_origin=B; + bool zero_string=false; + for (int j = 0; j < B_origin.chain.Nstrings; ++j) + for (int alpha = 0; alpha < B_origin.base.Nrap[j]; ++alpha) + if(abs(B_origin.lambda[j][alpha])* mu = new complex[sizeA]; + complex* lam = new complex[sizeB]; + int index=0; + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + { + mu[index]=(A.lambda[i][alpha] + 0.5 * -eta * (A.chain.Str_L[i] + 1.0 - 2.0 * a)); + index++; + } + +index=0; + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + { + lam[index]=(B.lambda[i][alpha] + 0.5 * -eta * (B.chain.Str_L[i] + 1.0 - 2.0 * a)); + index++; + } + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + +for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))); + + // Define regularized products in prefactors + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) + ln_prod3 += ln_Fn_F(A, j, alpha, a - 1); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + result += 2.0*real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) - A.lnnorm - B.lnnorm; +// Define the F ones earlier... + + int index_a = 0; + int index_b = 0; + complex Prod_powerN; + + //mu is the ground state! + //A -> mu, B -> lam +SQMat_CX H(0.0, A.base.Mdown); + index_a = 0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + complex Da; + complex Ca; + Da=eta/((mu[index_a]-eta*0.5)*(mu[index_a]+eta*0.5)); + Ca=eta*((mu[index_a]-eta*0.5)+(mu[index_a]+eta*0.5))/pow(((mu[index_a]-eta*0.5)*(mu[index_a]+eta*0.5)),2.0); + + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + + complex prodplus= complex(1.0,0.0); + complex prodminus= complex(1.0,0.0); + + // use simplified code for one-string here: original form of Hm2P matrix + prodplus = Fn_K (A, j, alpha, a, B, k, beta, 0);// 1.0/ (phi(mu[l]-lam[k]) phi(mu[l]-lam[k]+eta) ) + prodplus*= exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]+eta) / prod_l!=k |phi(lam[l]-lam[k]) |; + + prodminus = Fn_K (A, j, alpha, a, B, k, beta, 1);// 1.0/ (phi(mu[l]-lam[k]) phi(mu[l]-lam[k]-eta) ) + prodminus*= exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]-eta)/ prod_l!=k | phi(lam[l]-lam[k]) |; + + Prod_powerN = pow((B.lambda[k][beta] - eta*0.5) /(B.lambda[k][beta] + eta*0.5), complex (B.chain.Nsites)); + H[index_a][index_b] =eta*(prodplus-prodminus*Prod_powerN); + } // if (B.chain.Str_L == 1) + + else { + // */{ + + if (b > 1){ + H[index_a][index_b] = eta* Fn_K(A, j, alpha, a, B, k, beta, b-1)*exp(ln_Fn_G(A,B,k,beta,b-1))*exp(-real(ln_Fn_F(B, k, beta, b - 1)));//.../ prod_l!=k | phi(lam[l]-lam[k]) | + } + else if (b == 1) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + complex sum1 = complex(0.0,0.0); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) *exp(ln_FunctionG[0]+ ln_FunctionG[1] + - ln_FunctionF[0] - ln_FunctionF[1]);//sum term when i=0 + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp( ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); //sum term when i=n + + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) { + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + } + + H[index_a][index_b] = eta * exp(ln_FunctionF[0]+ ln_FunctionF[1] - ln_FunctionG[1]) * sum1 * exp( - real(ln_Fn_F(B, k, beta, b - 1))); //the absolute value prod_l!=k phi(lam[l]-lam[k]) : real(ln_...) + } // else if (b == B.chain.Str_L[k]) + } // else + index_b++; + }}} // sums over k, beta, b + + // now define the elements H[a][M] & H[a][M-1] + H[index_a][B.base.Mdown]=Da; + H[index_a][B.base.Mdown+1]=Ca; + index_a++; + }}} // sums over j, alpha, a + + complex logF=lndet_LU_CX_dstry(H); + result+=2.0*real(logF); + result+=log(2.0)-log((A.chain.Nsites-A.base.Mdown*2+3.0)*((A.chain.Nsites-A.base.Mdown*2+4.0))); + if (!(real_dev_conv) && abs(exp(result)-exp(prev_result))20){ + result=-300; + real_dev_conv=true; + } + + delete[] mu; + delete[] lam; + + } + //return(result); + return(0.5 * result); // Return ME, not MEsq +} + +} // namespace JSC diff --git a/src/HEIS/ln_Sz_ME_XXX.cc b/src/HEIS/ln_Sz_ME_XXX.cc new file mode 100644 index 0000000..aeeb41f --- /dev/null +++ b/src/HEIS/ln_Sz_ME_XXX.cc @@ -0,0 +1,247 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +***********************************************************/ + +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + +inline complex ln_Fn_F (XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) + ans += log(B.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex ln_Fn_G (XXX_Bethe_State& A, XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + ans += log(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex Fn_K (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return(1.0/((A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0)) ))); +} + +inline complex Fn_L (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return ((2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + )) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + +complex ln_Sz_ME (XXX_Bethe_State& A, XXX_Bethe_State& B) +{ + // This function returns the natural log of the S^z operator matrix element. + // The A and B states can contain strings. + + // Check that the two states refer to the same XXX_Chain + + if (A.chain != B.chain) JSCerror("Incompatible XXX_Chains in Sz matrix element."); + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown) JSCerror("Incompatible Mdown between the two states in Sz matrix element!"); + + //if (A.iK == B.iK && (A.label != B.label)) + if (A.iK == B.iK && (A.label.compare(B.label) != 0)) + return(-300.0); // matrix element identically vanishes + + // Some convenient arrays + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm((A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm((B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm((B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)))); + + // Define the F ones earlier... + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) + ln_prod3 += ln_Fn_F (A, j, alpha, a - 1); + + // ln_prod3 -= A.base.Mdown * log(abs(sin(A.chain.zeta))); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + // ln_prod4 -= B.base.Mdown * log(abs(sin(B.chain.zeta))); + + // Now proceed to build the Hm2P matrix + + SQMat_CX Hm2P(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + complex two_over_A_lambda_sq_plus_1over2sq; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + two_over_A_lambda_sq_plus_1over2sq = 2.0/((A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) * + (A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) + 0.25); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm2P matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta]); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta]); + + Prod_powerN = pow((B.lambda[k][beta] + 0.5 * II)/(B.lambda[k][beta] - 0.5 * II), complex (B.chain.Nsites)); + + Hm2P[index_a][index_b] = Fn_K_0_G_0 - Prod_powerN * Fn_K_1_G_2 + - two_over_A_lambda_sq_plus_1over2sq * exp(II*im_ln_Fn_F_B_0[k][beta]); + + } + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm2P[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + sum2 = 0.0; + + for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]]); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_Fn_F(B, k, beta, jsum - 1))); // include all string contributions F_B_0 in this term + + Hm2P[index_a][index_b] = prod_num * (sum1 - sum2 * two_over_A_lambda_sq_plus_1over2sq); + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + + index_a++; + }}} // sums over j, alpha, a + + //cout << "Matrix: " << endl; + //Hm2P.Print(); + + DP det = real(lndet_LU_CX_dstry(Hm2P)); + + complex ln_ME_sq = log(0.25 * A.chain.Nsites) + real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) + // + 2.0 * real(lndet_LU_CX_dstry(Hm2P)) + + 2.0 * det + - A.lnnorm - B.lnnorm; + + //cout << "ln_Sz: " << endl << ln_prod1 << "\t" << -ln_prod2 << "\t" << -ln_prod3 << "\t" << ln_prod4 << "\t" << 2.0 * det + // << "\t" << -A.lnnorm << "\t" << -B.lnnorm << endl; + + //return(ln_ME_sq); + return(0.5 * ln_ME_sq); // Return ME, not MEsq + +} + +} // namespace JSC diff --git a/src/HEIS/ln_Sz_ME_XXZ.cc b/src/HEIS/ln_Sz_ME_XXZ.cc new file mode 100644 index 0000000..d6d179e --- /dev/null +++ b/src/HEIS/ln_Sz_ME_XXZ.cc @@ -0,0 +1,563 @@ +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + + inline complex ln_Fn_F (XXZ_Bethe_State& B, int k, int beta, int b) + { + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + int arg = 0; + int absarg = 0; + int par_comb_1, par_comb_2; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + + par_comb_1 = B.chain.par[j] == B.chain.par[k] ? 1 : 0; + par_comb_2 = B.chain.par[k] == B.chain.par[j] ? 0 : B.chain.par[k]; + + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) { + + arg = B.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + absarg = abs(arg); + /* + prod_temp *= 0.5 * + ((B.sinhlambda[j][alpha] * B.coshlambda[k][beta] - B.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (B.chain.co_n_anis_over_2[absarg] * (1.0 + B.chain.par[j] * B.chain.par[k]) + - sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * (B.chain.par[k] - B.chain.par[j])) + + II * (B.coshlambda[j][alpha] * B.coshlambda[k][beta] - B.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) + * B.chain.si_n_anis_over_2[absarg] * (1.0 + B.chain.par[j] * B.chain.par[k]) + + B.chain.co_n_anis_over_2[absarg] * (B.chain.par[k] - B.chain.par[j])) ); + */ + prod_temp *= ((B.sinhlambda[j][alpha] * B.coshlambda[k][beta] - B.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (B.chain.co_n_anis_over_2[absarg] * par_comb_1 - sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * par_comb_2) + + II * (B.coshlambda[j][alpha] * B.coshlambda[k][beta] - B.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * par_comb_1 + B.chain.co_n_anis_over_2[absarg] * par_comb_2)); + + } + + if (counter++ > 100) { // we do at most 100 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + + }}} + + return(ans + log(prod_temp)); + } + + inline complex ln_Fn_G (XXZ_Bethe_State& A, XXZ_Bethe_State& B, int k, int beta, int b) + { + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + int arg = 0; + int absarg = 0; + int par_comb_1, par_comb_2; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + + par_comb_1 = A.chain.par[j] == B.chain.par[k] ? 1 : 0; + par_comb_2 = B.chain.par[k] == A.chain.par[j] ? 0 : B.chain.par[k]; + + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + arg = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + absarg = abs(arg); + + /* + prod_temp *= 0.5 * + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg] * (B.chain.par[k] - A.chain.par[j])) ); + */ + prod_temp *= ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg] * par_comb_1 - sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * par_comb_2) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * par_comb_1 + A.chain.co_n_anis_over_2[absarg] * par_comb_2)); + + if (counter++ > 100) { // we do at most 100 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + }}} + + return(ans + log(prod_temp)); + } + + inline complex Fn_K (XXZ_Bethe_State& A, int j, int alpha, int a, XXZ_Bethe_State& B, int k, int beta, int b) + { + int arg1 = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + int absarg1 = abs(arg1); + int arg2 = arg1 + 2; + int absarg2 = abs(arg2); + + return(4.0/( + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg1] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg1) * A.chain.si_n_anis_over_2[absarg1] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg1) * A.chain.si_n_anis_over_2[absarg1] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg1] * (B.chain.par[k] - A.chain.par[j])) ) + * + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg2] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg2) * A.chain.si_n_anis_over_2[absarg2] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg2) * A.chain.si_n_anis_over_2[absarg2] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg2] * (B.chain.par[k] - A.chain.par[j])) ) + )); + + } + + inline complex Fn_L (XXZ_Bethe_State& A, int j, int alpha, int a, XXZ_Bethe_State& B, int k, int beta, int b) + { + return (sinh(2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + + 0.25 * II * PI * complex(-A.chain.par[j] + B.chain.par[k]))) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); + } + + + // Version without phases: + complex ln_Sz_ME (XXZ_Bethe_State& A, XXZ_Bethe_State& B) + { + // This function returns the natural log of the S^z operator matrix element. + // The A and B states can contain strings. + + // Check that the two states refer to the same XXZ_Chain + + if (A.chain != B.chain) { + JSCerror("Incompatible XXZ_Chains in Sz matrix element."); + } + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown) { + cout << "Bra state: " << endl << A << endl; + cout << "Ket state: " << endl << B << endl; + cout << A.base.Mdown << "\t" << B.base.Mdown << endl; + JSCerror("Incompatible Mdown between the two states in Sz matrix element!"); + } + + if (A.iK == B.iK && (A.label != B.label)) + return(-300.0); // matrix element identically vanishes + + // Compute the sinh and cosh of rapidities + + A.Compute_sinhlambda(); + A.Compute_coshlambda(); + B.Compute_sinhlambda(); + B.Compute_coshlambda(); + + // Some convenient arrays + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(sinh(A.lambda[i][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - A.chain.par[i])))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - B.chain.par[i]))) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - B.chain.par[i])))); + + // Define the F ones earlier... + + complex ln_Fn_F_B_0; + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + ln_Fn_F_B_0 = ln_Fn_F(B, j, alpha, 0); + //re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + //im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F_B_0); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F_B_0); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + DP logabssinzeta = log(abs(sin(A.chain.anis))); + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + ln_prod3 += ln_Fn_F (A, j, alpha, a - 1); + } + + ln_prod3 -= A.base.Mdown * log(abs(sin(A.chain.anis))); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + ln_prod4 -= B.base.Mdown * log(abs(sin(B.chain.anis))); + + // Now proceed to build the Hm2P matrix + + SQMat_CX Hm2P(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + complex two_over_A_sinhlambda_sq_plus_sinzetaover2sq; + + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + two_over_A_sinhlambda_sq_plus_sinzetaover2sq = 2.0/((sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + * (sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + + pow(sin(0.5*A.chain.anis), 2.0)); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm2P matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinzeta); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinzeta); + + Prod_powerN = pow( B.chain.par[k] == 1 ? + (B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + : + (B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + , complex (B.chain.Nsites)); + + //cout << "Prod_powerN = " << Prod_powerN << "\t" << abs(Prod_powerN) << endl; + + Hm2P[index_a][index_b] = Fn_K_0_G_0 - Prod_powerN * Fn_K_1_G_2 - two_over_A_sinhlambda_sq_plus_sinzetaover2sq + * exp(II*im_ln_Fn_F_B_0[k][beta] + logabssinzeta); + + } + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm2P[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + sum2 = 0.0; + + for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]] + logabssinzeta); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_FunctionF[jsum - 1]) + logabssinzeta); + // include all string contributions F_B_0 in this term + + Hm2P[index_a][index_b] = prod_num * (sum1 - sum2 * two_over_A_sinhlambda_sq_plus_sinzetaover2sq); + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + index_a++; + }}} // sums over j, alpha, a + + DP re_ln_det = real(lndet_LU_CX_dstry(Hm2P)); + + complex ln_ME_sq = log(0.25 * A.chain.Nsites) + real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) + + 2.0 * /*real(lndet_LU_CX_dstry(Hm2P))*/ re_ln_det - A.lnnorm - B.lnnorm; + + //cout << endl << ln_prod1 << "\t" << ln_prod2 << "\t" << ln_prod3 << "\t" << ln_prod4 << "\t" << A.lnnorm << "\t" << B.lnnorm + //<< "\t" << re_ln_det << "\t" << ln_ME_sq << endl; + + //return(ln_ME_sq); + return(0.5 * ln_ME_sq); // Return ME, not MEsq + + } + + // Version with phases: + complex ln_Sz_ME_with_phase (XXZ_Bethe_State& A, XXZ_Bethe_State& B) + { + // This function returns the natural log of the S^z operator matrix element. + // The A and B states can contain strings. + + // Check that the two states refer to the same XXZ_Chain + + if (A.chain != B.chain) JSCerror("Incompatible XXZ_Chains in Sz matrix element."); + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown) { + cout << "Bra state: " << endl << A << endl; + cout << "Ket state: " << endl << B << endl; + cout << A.base.Mdown << "\t" << B.base.Mdown << endl; + JSCerror("Incompatible Mdown between the two states in Sz matrix element!"); + } + + if (A.iK == B.iK && (A.label != B.label)) + return(-300.0); // matrix element identically vanishes + + // Compute the sinh and cosh of rapidities + + A.Compute_sinhlambda(); + A.Compute_coshlambda(); + B.Compute_sinhlambda(); + B.Compute_coshlambda(); + + // Some convenient arrays + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(sinh(A.lambda[i][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - A.chain.par[i]))); + + complex shB; + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(shB = sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - B.chain.par[i]))) > 100.0 * MACHINE_EPS_SQ) + //ln_prod2 += log(norm(sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + // + 0.25 * II * PI * (1.0 - B.chain.par[i])))); + ln_prod2 += log(shB); + + // Define the F ones earlier... + + complex ln_Fn_F_B_0, ln_Fn_G_0, ln_Fn_G_2; + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + //re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + //im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + ln_Fn_F_B_0 = ln_Fn_F(B, j, alpha, 0); + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F_B_0); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F_B_0); + //re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + //im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + ln_Fn_G_0 = ln_Fn_G(A, B, j, alpha, 0); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G_0); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G_0); + //re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + //im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + ln_Fn_G_0 = ln_Fn_G(A, B, j, alpha, 2); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G_2); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G_2); + } + } + + DP logsinzeta = log(sin(A.chain.anis)); + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + ln_prod3 += ln_Fn_F (A, j, alpha, a - 1); + } + + //ln_prod3 -= A.base.Mdown * log(abs(sin(A.chain.anis))); + ln_prod3 -= A.base.Mdown * logsinzeta; + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + //ln_prod4 -= B.base.Mdown * log(abs(sin(B.chain.anis))); + ln_prod4 -= B.base.Mdown * logsinzeta; + + // Now proceed to build the Hm2P matrix + + SQMat_CX Hm2P(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + complex two_over_A_sinhlambda_sq_plus_sinzetaover2sq; + + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + two_over_A_sinhlambda_sq_plus_sinzetaover2sq = 2.0/((sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + * (sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + + pow(sin(0.5*A.chain.anis), 2.0)); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm2P matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta] + logsinzeta); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta] + logsinzeta); + + Prod_powerN = pow( B.chain.par[k] == 1 ? + (B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + : + (B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + , complex (B.chain.Nsites)); + + //cout << "Prod_powerN = " << Prod_powerN << "\t" << abs(Prod_powerN) << endl; + + Hm2P[index_a][index_b] = Fn_K_0_G_0 - Prod_powerN * Fn_K_1_G_2 - two_over_A_sinhlambda_sq_plus_sinzetaover2sq + * exp(II*im_ln_Fn_F_B_0[k][beta] + logsinzeta); + + } + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm2P[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + sum2 = 0.0; + + for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]] + logsinzeta); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_FunctionF[jsum - 1]) + logsinzeta); + // include all string contributions F_B_0 in this term + + Hm2P[index_a][index_b] = prod_num * (sum1 - sum2 * two_over_A_sinhlambda_sq_plus_sinzetaover2sq); + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + index_a++; + }}} // sums over j, alpha, a + + complex ln_det = lndet_LU_CX_dstry(Hm2P); + + complex ln_ME = log(0.5 * sqrt(A.chain.Nsites)) + ln_prod1 - ln_prod2 - ln_prod3 + ln_prod4 + 2.0 * ln_det - A.lnnorm - B.lnnorm; + + //cout << endl << ln_prod1 << "\t" << ln_prod2 << "\t" << ln_prod3 << "\t" << ln_prod4 << "\t" << A.lnnorm << "\t" << B.lnnorm + //<< "\t" << re_ln_det << "\t" << ln_form_factor_sq << endl; + + //return(ln_ME_sq); + return(ln_ME); // Return ME, not MEsq + + } + +} // namespace JSC diff --git a/src/HEIS/ln_Sz_ME_XXZ_gpd.cc b/src/HEIS/ln_Sz_ME_XXZ_gpd.cc new file mode 100644 index 0000000..2f9eefb --- /dev/null +++ b/src/HEIS/ln_Sz_ME_XXZ_gpd.cc @@ -0,0 +1,304 @@ +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + +inline complex ln_Fn_F (XXZ_gpd_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + /* + if (!((j == k) && (alpha == beta) && (a == b))) + ans += log(-II * sin(B.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.eta * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b)))); + */ + + if (!((j == k) && (alpha == beta) && (a == b))) { + + // arg = B.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + // absarg = abs(arg); + + prod_temp *= -II * (B.sinlambda[j][alpha] * B.coslambda[k][beta] - B.coslambda[j][alpha] * B.sinlambda[k][beta]) + * cosh(0.5 * B.chain.anis * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + + (B.coslambda[j][alpha] * B.coslambda[k][beta] + B.sinlambda[j][alpha] * B.sinlambda[k][beta]) + * sinh(0.5 * B.chain.anis * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + + } + + + if (counter++ > 10) { // we do at most 10 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + + }}} + + // return(ans); + return(ans + log(prod_temp)); +} + +inline complex ln_Fn_G (XXZ_gpd_Bethe_State& A, XXZ_gpd_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + /* + prod_temp *= -II * sin(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.eta * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + */ + + // arg = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + // absarg = abs(arg); + + prod_temp *= -II * (A.sinlambda[j][alpha] * B.coslambda[k][beta] - A.coslambda[j][alpha] * B.sinlambda[k][beta]) + * cosh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + + (A.coslambda[j][alpha] * B.coslambda[k][beta] + A.sinlambda[j][alpha] * B.sinlambda[k][beta]) + * sinh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + + if (counter++ > 25) { // we do at most 25 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + }}} + + // return(ans); + return(ans + log(prod_temp)); +} + +inline complex Fn_K (XXZ_gpd_Bethe_State& A, int j, int alpha, int a, XXZ_gpd_Bethe_State& B, int k, int beta, int b) +{ + /* + return(-1.0/(sin(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.zeta * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + * sin(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.zeta * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0))))); + */ + + // int arg1 = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + // int absarg1 = abs(arg1); + // int arg2 = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b - 1); + // int arg2 = arg1 + 2; + // int absarg2 = abs(arg2); + + return(1.0/(( -II * (A.sinlambda[j][alpha] * B.coslambda[k][beta] - A.coslambda[j][alpha] * B.sinlambda[k][beta]) + * cosh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + + (A.coslambda[j][alpha] * B.coslambda[k][beta] + A.sinlambda[j][alpha] * B.sinlambda[k][beta]) + * sinh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b)))) * + (-II * (A.sinlambda[j][alpha] * B.coslambda[k][beta] - A.coslambda[j][alpha] * B.sinlambda[k][beta]) + * cosh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0))) + + (A.coslambda[j][alpha] * B.coslambda[k][beta] + A.sinlambda[j][alpha] * B.sinlambda[k][beta]) + * sinh(0.5 * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0)))))); + +} + +inline complex Fn_L (XXZ_gpd_Bethe_State& A, int j, int alpha, int a, XXZ_gpd_Bethe_State& B, int k, int beta, int b) +{ + /* + complex ans = -II * sin(2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.zeta * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)))); + + return (ans * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); + */ + return (-II * sin(2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)))) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + +complex ln_Sz_ME (XXZ_gpd_Bethe_State& A, XXZ_gpd_Bethe_State& B) +{ + // This function returns the natural log of the S^z operator matrix element. + // The A and B states can contain strings. + + // Check that the two states refer to the same XXZ_gpd_Chain + + if (A.chain != B.chain) JSCerror("Incompatible XXZ_gpd_Chains in Sz matrix element."); + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown) JSCerror("Incompatible Mdown between the two states in Sz matrix element!"); + + if (A.iK == B.iK && (A.label != B.label)) + return(-300.0); // matrix element identically vanishes + + // Compute the sin and cos of rapidities + + A.Compute_sinlambda(); + A.Compute_coslambda(); + B.Compute_sinlambda(); + B.Compute_coslambda(); + + // Some convenient arrays + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(sin(A.lambda[i][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(sin(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(sin(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)))); + + // Define the F ones earlier... + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + DP logabssinheta = log(abs(sinh(A.chain.anis))); + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) + ln_prod3 += ln_Fn_F(A, j, alpha, a - 1); + + ln_prod3 -= A.base.Mdown * log(abs(sinh(A.chain.anis))); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + ln_prod4 -= B.base.Mdown * log(abs(sinh(B.chain.anis))); + + // Now proceed to build the Hm2P matrix + + SQMat_CX Hm2P(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + complex two_over_A_sinlambda_sq_plus_sinhetaover2sq = 0.0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + two_over_A_sinlambda_sq_plus_sinhetaover2sq = -2.0/((sin(A.lambda[j][alpha] // minus sign: from 1/(sinh^2... to -1/(sin^2... + + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a))) * + (sin(A.lambda[j][alpha] // minus sign: from 1/(sinh^2... to -1/(sin^2... + + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a))) + + pow(sinh(0.5*A.chain.anis), 2.0)); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm2P matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinheta); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinheta); + + Prod_powerN = pow((B.sinlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.coslambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.sinlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.coslambda[k][beta] * B.chain.si_n_anis_over_2[1]), + complex (B.chain.Nsites)); + + Hm2P[index_a][index_b] = Fn_K_0_G_0 - Prod_powerN * Fn_K_1_G_2 + - two_over_A_sinlambda_sq_plus_sinhetaover2sq * exp(II*im_ln_Fn_F_B_0[k][beta] + logabssinheta); + + } + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm2P[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + sum2 = 0.0; + + for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]] + logabssinheta); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_Fn_F(B, k, beta, jsum - 1)) + logabssinheta); + // include all string contributions F_B_0 in this term + + Hm2P[index_a][index_b] = prod_num * (sum1 - sum2 * two_over_A_sinlambda_sq_plus_sinhetaover2sq); + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + index_a++; + }}} // sums over j, alpha, a + + complex ln_ME_sq = log(0.25 * A.chain.Nsites) + real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) + + 2.0 * real(lndet_LU_CX_dstry(Hm2P)) - A.lnnorm - B.lnnorm; + + //cout << endl << ln_prod1 << "\t" << ln_prod2 << "\t" << ln_prod3 << "\t" << ln_prod4 << "\t" << A.lnnorm << "\t" << B.lnnorm << "\t" + // << lndet_LU_CX(Hm2P) << endl; + + //return(ln_ME_sq); + return(0.5 * ln_ME_sq); // Return ME, not MEsq + +} + +} // namespace JSC diff --git a/src/HEIS/ln_Szm_p_Smz_ME_XXX.cc b/src/HEIS/ln_Szm_p_Smz_ME_XXX.cc new file mode 100644 index 0000000..9cb8f6f --- /dev/null +++ b/src/HEIS/ln_Szm_p_Smz_ME_XXX.cc @@ -0,0 +1,502 @@ +#include "JSC.h" + +using namespace JSC; + + +namespace JSC { + inline complex phi(complex x){return x;} +inline complex a(complex x){return 1;} +inline complex b(complex x,complex y, complex eta){ return phi(x-y)/phi(x-y+complex(0.0,1.0)*eta);} +inline complex d(complex x, complex xi, complex eta, int N){return pow(b(x,xi,eta),N);} + + +inline complex ln_Fn_F (XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) + ans += log(B.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex ln_Fn_G (XXX_Bethe_State& A, XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + ans += log(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex Fn_K (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return(1.0/((A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0)) ))); +} + +inline complex Fn_L (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return ((2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + )) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + + +complex ln_Szm_p_Smz_ME (XXX_Bethe_State& A, XXX_Bethe_State& B) +{ + //clock_t start_time_local = clock(); + //A has to be the ground state! + + // This function returns the natural log of the S^z operator matrix element. + // The A and B states can contain strings. + + // Check that the two states refer to the same XXX_Chain + + if (A.chain != B.chain) JSCerror("Incompatible XXX_Chains in Szm_p_Smz matrix element."); + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown + 1) JSCerror("Incompatible Mdown between the two states in SzSm_p_SmSz matrix element!"); + + //if (B.type_id > 999999LL) return(complex(-300.0)); + + //add a delta to the origin of the centered strings + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + if(abs(A.lambda[i][alpha])<5.55112e-12)A.lambda[i][alpha]=5.55112e-12; + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + if(abs(B.lambda[i][alpha])<5.55112e-5)B.lambda[i][alpha]=5.55112e-5; + + // Some convenient arrays + + complex i=complex(0.0,1.0); + complex eta=-i; + complex ln_prod = complex(0.0,0.0); + complex result; + result=log(B.chain.Nsites*1.0/4.0); + + int sizeA=0; +int sizeB=0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + sizeA+=A.base.Nrap[i]*A.chain.Str_L[i]; + for (int i = 0; i < B.chain.Nstrings; ++i) + sizeB+=B.base.Nrap[i]*B.chain.Str_L[i]; + + complex* mu = new complex[sizeA]; + complex* lam = new complex[sizeB]; + int index=0; + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + { + // mu[index]=(A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a)); + mu[index]=(A.lambda[i][alpha] + 0.5 * -eta * (A.chain.Str_L[i] + 1.0 - 2.0 * a)); + index++; + } + +index=0; + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + { + // lam[index]=(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a)); + lam[index]=(B.lambda[i][alpha] + 0.5 * -eta * (B.chain.Str_L[i] + 1.0 - 2.0 * a)); + index++; + } + + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))); + + // Define the F ones earlier... + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + + +// for (int i = 0; i < A.chain.Nstrings; ++i) +// for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) +// for (int a = 1; a <= A.chain.Str_L[i]; ++a) +// ln_prod+=log(abs(phi((A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a))+eta*0.5))) + +// for (int i = 0; i < B.chain.Nstrings; ++i) +// for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) +// for (int a = 1; a <= B.chain.Str_L[i]; ++a) +// ln_prod+=-log(abs(phi(-(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a))+eta*0.5))); + +// for (int i = 0; i < B.chain.Nstrings; ++i) +// for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) +// for (int a = 1; a <= B.chain.Str_L[i]; ++a) +// for (int j = 0; j < B.chain.Nstrings; ++j) +// for (int beta = 0; beta < B.base.Nrap[j]; ++beta) +// for (int b = 1; b <= B.chain.Str_L[j]; ++b) +// if(i!=j || alpha!=beta ||a!=b)ln_prod3+=-0.5*log(abs(phi((B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a))-(B.lambda[j][beta] + 0.5 * II * (B.chain.Str_L[j] + 1.0 - 2.0 * b))-eta))); + +// for (int i = 0; i < A.chain.Nstrings; ++i) +// for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) +// for (int a = 1; a <= A.chain.Str_L[i]; ++a) +// for (int j = 0; j < A.chain.Nstrings; ++j) +// for (int beta = 0; beta < A.base.Nrap[j]; ++beta) +// for (int b = 1; b <= A.chain.Str_L[j]; ++b) +// if(abs((A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a))-(A.lambda[j][beta] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * b))-eta)!=0 && (i!=j && alpha!=beta && a!=b)) +// ln_prod3+=-0.5*log(abs((A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a))-(A.lambda[j][beta] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * b))-eta)); + + + +//mu is the ground state! +//A -> mu, B -> lam +complex prod=complex(1.0,0.0); +complex prod2=complex(1.0,0.0); +complex prod3=complex(1.0,0.0); + for(int l=0; l H[sizeA][sizeA]; + // cout<<"mu[l]:"; +// for(int l=0; l Prod_powerN; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + complex prodplus= complex(1.0,0.0); + complex prodminus= complex(1.0,0.0); + + // use simplified code for one-string here: original form of Hm2P matrix + + prodplus = Fn_K (A, j, alpha, a, B, k, beta, 0);// 1.0/ (phi(mu[l]-lam[k]) phi(mu[l]-lam[k]+eta) ) + prodplus*= exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]+eta) / prod_l!=k |phi(lam[l]-lam[k]) |; + prodminus = Fn_K (A, j, alpha, a, B, k, beta, 1);// 1.0/ (phi(mu[l]-lam[k]) phi(mu[l]-lam[k]-eta) ) + prodminus*= exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]-eta)/ prod_l!=k | phi(lam[l]-lam[k]) |; + + Prod_powerN = pow((B.lambda[k][beta] -eta*0.5) /(B.lambda[k][beta] +eta*0.5), complex (B.chain.Nsites)); + + Hm[index_a][index_b] =eta*(prodplus-prodminus*Prod_powerN); + + } // if (B.chain.Str_L == 1) + + else { + + if (b > 1) Hm[index_a][index_b] = eta* Fn_K(A, j, alpha, a, B, k, beta, b-1)*exp(ln_Fn_G(A,B,k,beta,b-1))*exp(-real(ln_Fn_F(B, k, beta, b - 1)));//.../ prod_l!=k | phi(lam[l]-lam[k]) | + else if (b == 1) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + complex sum1 = complex(0.0,0.0); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) *exp(ln_FunctionG[0]+ ln_FunctionG[1] //sum term when i=0 + - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) //sum term when i=n + * exp( ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + + Hm[index_a][index_b] = eta * exp(ln_FunctionF[0]+ ln_FunctionF[1] - ln_FunctionG[1]) * sum1 * exp( - real(ln_Fn_F(B, k, beta, b - 1))); //the absolute value prod_l!=k phi(lam[l]-lam[k]) : real(ln_...) + } // else if (b == B.chain.Str_L[k]) + } // else + index_b++; + }}} // sums over k, beta, b + + // now define the elements Hm[a][M] + + //Hm[index_a][B.base.Mdown] = one_over_A_lambda_sq_plus_1over2sq; + //Hm[index_a][B.base.Mdown] = eta/((A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) * + // (A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) + 0.25); + Hm[index_a][B.base.Mdown] = eta/((mu[index_a]-eta*0.5)*(mu[index_a]+eta*0.5)); + index_a++; + }}} // sums over j, alpha, a + +// cout<<"Hm:"; +// for(int j=0; j F= complex(0.0,0.0); + + complex detmatrix; + detmatrix=exp(lndet_LU_CX_dstry(Hm)); + //cout<<"exp(lndet_LU_CX(Hm)):"< mu, B -> lam +SQMat_CX G(0.0, A.base.Mdown); +SQMat_CX BbDa(0.0, A.base.Mdown); + index_a = 0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + complex Da; + complex Ca; + Da=eta/((mu[index_a]-eta*0.5)*(mu[index_a]+eta*0.5)); + Ca=eta*((mu[index_a]-eta*0.5)+(mu[index_a]+eta*0.5))/pow(((mu[index_a]-eta*0.5)*(mu[index_a]+eta*0.5)),2.0); + + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + /*if (B.chain.Str_L[k] == 1) { + + complex Bb; + Bb=-(phi(lam[index_b]+eta*0.5)+phi(lam[index_b]-eta*0.5)); + complex product=complex(1.0,0.0); + for(int o=0; o prodplus= complex(1.0,0.0); + complex prodminus= complex(1.0,0.0); + + // use simplified code for one-string here: original form of Hm2P matrix + prodplus = Fn_K (A, j, alpha, a, B, k, beta, 0);// 1.0/ (phi(mu[l]-lam[k]) phi(mu[l]-lam[k]+eta) ) + prodplus*= exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]+eta) / prod_l!=k |phi(lam[l]-lam[k]) |; + //prodplus*= exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta]);// - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]+eta) / prod_l!=k |phi(lam[l]-lam[k]) |; + prodminus = Fn_K (A, j, alpha, a, B, k, beta, 1);// 1.0/ (phi(mu[l]-lam[k]) phi(mu[l]-lam[k]-eta) ) + prodminus*= exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]-eta)/ prod_l!=k | phi(lam[l]-lam[k]) |; + // prodminus*= exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta]);// - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]-eta)/ prod_l!=k | phi(lam[l]-lam[k]) |; + + Prod_powerN = pow((B.lambda[k][beta] - eta*0.5) /(B.lambda[k][beta] + eta*0.5), complex (B.chain.Nsites)); + + G[index_a][index_b] =eta*(prodplus-prodminus*Prod_powerN); + BbDa[index_a][index_b]=Bb*Da*exp(- re_ln_Fn_F_B_0[k][beta]); + + } // if (B.chain.Str_L == 1) + + else { + */{ + + if (b > 1){ + G[index_a][index_b] = eta* Fn_K(A, j, alpha, a, B, k, beta, b-1)*exp(ln_Fn_G(A,B,k,beta,b-1))*exp(-real(ln_Fn_F(B, k, beta, b - 1)));//.../ prod_l!=k | phi(lam[l]-lam[k]) | + BbDa[index_a][index_b] =0 ; + } + else if (b == 1) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + complex sum1 = complex(0.0,0.0); + complex sum2 = complex(0.0,0.0); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) *exp(ln_FunctionG[0]+ ln_FunctionG[1] + - ln_FunctionF[0] - ln_FunctionF[1]);//sum term when i=0 + //sum2 doesn't have a i=0 term + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp( ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); //sum term when i=n + sum2 += exp(ln_FunctionG[B.chain.Str_L[k]]- ln_FunctionF[B.chain.Str_L[k]] ) + * (phi((B.lambda[k][beta] + 0.5 * II * (B.chain.Str_L[k] + 1.0 - 2.0 * B.chain.Str_L[k]))-eta*0.5) + + phi((B.lambda[k][beta] + 0.5 * II * (B.chain.Str_L[k] + 1.0 - 2.0 * B.chain.Str_L[k]))+eta*0.5) ); //sum term when i=n + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) { + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + sum2 += exp(ln_FunctionG[jsum]- ln_FunctionF[jsum] ) + * (phi((B.lambda[k][beta] + 0.5 * II * (B.chain.Str_L[k] + 1.0 - 2.0 * jsum))-eta*0.5) + + phi((B.lambda[k][beta] + 0.5 * II * (B.chain.Str_L[k] + 1.0 - 2.0 * jsum))+eta*0.5) ); + } + + G[index_a][index_b] = eta * exp(ln_FunctionF[0]+ ln_FunctionF[1] - ln_FunctionG[1]) * sum1 * exp( - real(ln_Fn_F(B, k, beta, b - 1))); //the absolute value prod_l!=k phi(lam[l]-lam[k]) : real(ln_...) + BbDa[index_a][index_b] = - Da* exp(ln_FunctionF[0]+ ln_FunctionF[1] - ln_FunctionG[1]) * sum2 * exp( - real(ln_Fn_F(B, k, beta, b - 1)));//the absolute value prod_l!=k phi(lam[l]-lam[k]) : real(ln_...) + } // else if (b == B.chain.Str_L[k]) + } // else + index_b++; + }}} // sums over k, beta, b + + // now define the elements Hm[a][M] + + + //Hm[index_a][B.base.Mdown] = eta/((A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) * (A.lambda[j][alpha] + 0.5 * II * (A.chain.Str_L[j] + 1.0 - 2.0 * a)) + 0.25); + G[index_a][B.base.Mdown]=Ca; + BbDa[index_a][B.base.Mdown]=0; + index_a++; + }}} // sums over j, alpha, a + +SQMat_CX matrix(0.0, A.base.Mdown); + for(int a=0; a sum=prod*(exp(lndet_LU_CX_dstry(matrix))-exp(lndet_LU_CX_dstry(G))); + //sum=conj(prod)*(exp(lndet_LU_CX_dstry(matrix))-exp(lndet_LU_CX_dstry(G))); + + //cout<<"an(n):"<< an(n)<< "det(matrix):"< F2=exp(eta*(B.K-A.K))*(-2.0*sum); + complex F3=(exp(eta*(B.K-A.K)))*(detmatrix); + F=detmatrix; + // cout<<"F1:"< ln_ME_sq = result; + +// if (real(ln_ME_sq) > 10.0) ln_ME_sq = complex (-300.0); // fix for artificial divergences + + delete[] mu; + delete[] lam; + + +//return(ln_ME_sq); + return(0.5 * ln_ME_sq); // Return ME, not MEsq + +} + +} // namespace JSC diff --git a/src/HEIS/ln_Szz_ME_XXX.cc b/src/HEIS/ln_Szz_ME_XXX.cc new file mode 100644 index 0000000..830beea --- /dev/null +++ b/src/HEIS/ln_Szz_ME_XXX.cc @@ -0,0 +1,417 @@ +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + +complex phi(complex x){return x;} +complex a(complex x){return 1;} +complex b(complex x,complex y, complex eta){ return phi(x-y)/phi(x-y+complex(0.0,1.0)*eta);} +complex d(complex x, complex xi, complex eta, int N){return pow(b(x,xi,eta),N);} + + +inline complex ln_Fn_F (XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) + ans += log(B.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (B.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex ln_Fn_G (XXX_Bethe_State& A, XXX_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + ans += log(A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))); + } + } + } + + return(ans); +} + +inline complex Fn_K (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return(1.0/((A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b))) + * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 1.0)) ))); +} + +inline complex Fn_L (XXX_Bethe_State& A, int j, int alpha, int a, XXX_Bethe_State& B, int k, int beta, int b) +{ + return ((2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + )) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + + complex ln_Szz_ME (XXX_Bethe_State& A, XXX_Bethe_State& B){ + //clock_t start_time_local = clock(); + //B has to be the ground state! + const DP Zero_Center_Thres=1.0e-5; + const DP real_dev=1.0e-14; + const DP Diff_ME_Thres=1.e-6; + + // This function returns the natural log of the S^z operator matrix element. + // The A and B states can contain strings. + + // Check that the two states refer to the same XXX_Chain + + if (A.chain != B.chain) JSCerror("Incompatible XXX_Chains in Szz matrix element."); + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown) JSCerror("Incompatible Mdown between the two states in Szz matrix element!"); + + complex eta=-II; + complex ln_prod = complex(0.0,0.0); + complex result=-300; + complex prev_result=-300; + + //if(A.String_delta()> HEIS_deltaprec) return(complex(-300.0)); // DEPRECATED in ++T_9 + +if((A.E)==(B.E) && B.chain.Nsites ==B.base.Mdown*2){ + return(2*log(abs((B.E+(B.chain.Nsites)/4.0)/(3.0*sqrt(B.chain.Nsites))))); +} + + XXX_Bethe_State A_origin; A_origin=A; + bool zero_string=false; + for (int j = 0; j < A_origin.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A_origin.base.Nrap[j]; ++alpha) + if(abs(A_origin.lambda[j][alpha])* mu = new complex[sizeA]; + complex* lam = new complex[sizeB]; + int index=0; + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + { + mu[index]=(A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a)); + index++; + } + +index=0; + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + { + lam[index]=(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a)); + index++; + } + + Lambda re_ln_Fn_F_A_0(A.chain, A.base); + Lambda im_ln_Fn_F_A_0(A.chain, A.base); + Lambda re_ln_Fn_F_B_0(B.chain, B.base); + Lambda im_ln_Fn_F_B_0(B.chain, B.base); + Lambda re_ln_Fn_G_0(B.chain, B.base); + Lambda im_ln_Fn_G_0(B.chain, B.base); + Lambda re_ln_Fn_G_2(B.chain, B.base); + Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(A.lambda[i][alpha] + 0.5 * II * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0)) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(B.lambda[i][alpha] + 0.5 * II * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0))); + + // Define the F ones earlier... + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_A_0[j][alpha] = real(ln_Fn_F(A, j, alpha, 0)); + im_ln_Fn_F_A_0[j][alpha] = imag(ln_Fn_F(A, j, alpha, 0)); + } + } + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + // Define regularized products in prefactors + +for (int k = 0; k < A.chain.Nstrings; ++k) + for (int beta = 0; beta < A.base.Nrap[k]; ++beta) + for (int b = 1; b <= A.chain.Str_L[k]; ++b) { + if (b == 1)ln_prod3 += re_ln_Fn_F_A_0[k][beta]; + else if (b > 1) ln_prod3 += ln_Fn_F(A, k, beta, b - 1);/*TEST*/ + } + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + +complex prod=complex(1.0,0.0); + for(int l=0; l Prod_powerN; + + +SQMat_CX H(0.0, A.base.Mdown); +SQMat_CX P(0.0, A.base.Mdown); + index_a = 0; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + complex Da; + complex Ca; + Da=eta/((mu[index_a]-eta*0.5)*(mu[index_a]+eta*0.5)); + + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + complex Bb=complex(1.0,0.0); + for(int o=0; o prodplus= complex(1.0,0.0); + complex prodminus= complex(1.0,0.0); + + // use simplified code for one-string here: original form of Hm2P matrix + prodplus = Fn_K (A, j, alpha, a, B, k, beta, 0); + + prodplus*= exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta]); + + prodminus = Fn_K (A, j, alpha, a, B, k, beta, 1); + + prodminus*= exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta]); + + Prod_powerN = pow((B.lambda[k][beta] - eta*0.5) /(B.lambda[k][beta] + eta*0.5), complex (B.chain.Nsites)); + + H[index_a][index_b] = eta*(prodplus-prodminus*Prod_powerN); + P[index_a][index_b] = Bb*Da*exp(- re_ln_Fn_F_B_0[k][beta]); + + } // if (B.chain.Str_L == 1) + + else { + + if (b > 1){ + H[index_a][index_b] = eta* Fn_K(A, j, alpha, a, B, k, beta, b-1)*exp(ln_Fn_G(A,B,k,beta,b-1))*exp(-real(ln_Fn_F(B, k, beta, b - 1)));//.../ prod_l!=k | phi(lam[l]-lam[k]) | + P[index_a][index_b] =0 ; + } + else if (b == 1) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + complex sum1 = complex(0.0,0.0); + complex sum2 = complex(0.0,0.0); + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) *exp(ln_FunctionG[0]+ ln_FunctionG[1] + - ln_FunctionF[0] - ln_FunctionF[1]);//sum term when i=0 + //sum2 doesn't have a i=0 term + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp( ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); //sum term when i=n + sum2 += exp(ln_FunctionG[B.chain.Str_L[k]]- ln_FunctionF[B.chain.Str_L[k]] ); //sum term when i=n + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) { + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + sum2 += exp(ln_FunctionG[jsum]- ln_FunctionF[jsum] ); + } + + H[index_a][index_b] = eta * exp(ln_FunctionF[0]+ ln_FunctionF[1] - ln_FunctionG[1]) * sum1 * exp( - real(ln_Fn_F(B, k, beta, b - 1))); //the absolute value prod_l!=k phi(lam[l]-lam[k]) : real(ln_...) + P[index_a][index_b] = - Da* exp(ln_FunctionF[0]+ ln_FunctionF[1] - ln_FunctionG[1]) * sum2 * exp( - real(ln_Fn_F(B, k, beta, b - 1)));//the absolute value prod_l!=k phi(lam[l]-lam[k]) : real(ln_...) + } // else if (b == B.chain.Str_L[k]) + } // else + index_b++; + }}} // sums over k, beta, b + + index_a++; + }}} // sums over j, alpha, a + +complex F= complex(0.0,0.0); + + +SQMat_CX matrix(0.0, A.base.Mdown); +for(int j=0; j detmatrix; + detmatrix=exp(lndet_LU_CX_dstry(matrix)+0.5*factor); + +SQMat_CX Gn(0.0, A.base.Mdown); + + complex sum_n=complex(0.0,0.0); + for(int n=0; n An; + An=-phi(lam[n]-eta*0.5); + for(int m=0; m Da; + complex Ca; + Da=eta/((mu[index_a]-eta*0.5)*(mu[index_a]+eta*0.5)); + Ca=(eta)*((mu[index_a]-eta*0.5)+(mu[index_a]+eta*0.5))/pow(((mu[index_a]-eta*0.5)*(mu[index_a]+eta*0.5)),2.0); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if(index_b==n){ + Gn[index_a][index_b] = Ca*exp(-real(ln_Fn_F(B, k, beta, b - 1))); + BnbDa[index_a][index_b] = 0; + } // else (index_b!=n) + else if (B.chain.Str_L[k] == 1) { + complex Bnb; + Bnb=-phi(lam[index_b]+eta*0.5)/phi(lam[n]-lam[index_b]-eta); + complex product=complex(1.0,0.0); + for(int o=0; o prodplus= complex(1.0,0.0); + complex prodminus= complex(1.0,0.0); + + // use simplified code for one-string here: original form of Hm2P matrix + prodplus = Fn_K (A, j, alpha, a, B, k, beta, 0);// 1.0/ (phi(mu[l]-lam[k]) phi(mu[l]-lam[k]+eta) ) + prodplus*= exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]+eta) / prod_l!=k |phi(lam[l]-lam[k]) |; + prodminus = Fn_K (A, j, alpha, a, B, k, beta, 1);// 1.0/ (phi(mu[l]-lam[k]) phi(mu[l]-lam[k]-eta) ) + prodminus*= exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta]);//Prod phi(mu[l]-lam[k]-eta)/ prod_l!=k | phi(lam[l]-lam[k]) |; + + Prod_powerN = pow((B.lambda[k][beta] - eta*0.5) /(B.lambda[k][beta] + eta*0.5), complex (B.chain.Nsites)); + + Gn[index_a][index_b] =eta*(prodplus-prodminus*Prod_powerN); + BnbDa[index_a][index_b]=Bnb*Da*exp(- re_ln_Fn_F_B_0[k][beta]); + + } // if (B.chain.Str_L == 1) + + else{ + JSCerror("The Szz matrix element computation is not able to handle string states in the B.state (second argument). This is in development..."); + + } // else + index_b++; + }}} // sums over k, beta, b + + index_a++; + }}} // sums over j, alpha, a + +SQMat_CX matrix(0.0, A.base.Mdown); + for(int a=0; a20){ + result=-300; + real_dev_conv=true; + } + + delete[] mu; + delete[] lam; + + } + + //return(result); + return(0.5 * result); // Return ME, not MEsq +} + +} // namespace JSC diff --git a/src/INFPREC/Infprec.cc b/src/INFPREC/Infprec.cc new file mode 100644 index 0000000..fd64016 --- /dev/null +++ b/src/INFPREC/Infprec.cc @@ -0,0 +1,25 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Infprec.cc + +Purpose: Definitions for infinite precision arithmetic classes. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + //*************************************************************************************************** + + + +} // namespace JSC diff --git a/src/INTEG/Integration.cc b/src/INTEG/Integration.cc new file mode 100644 index 0000000..663ee05 --- /dev/null +++ b/src/INTEG/Integration.cc @@ -0,0 +1,1648 @@ +/**************************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +Integration.cc + +Defines all functions to perform integration of functions. + + +******************************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + // ********************* class I_table ****************************** + + I_table::I_table (DP (*function_ref) (DP, DP), DP rhomin_ref, DP rhomax_ref, int Nvals_ref, DP req_prec) { + + function = function_ref; + + if (!(*this).Load(rhomin_ref, rhomax_ref, Nvals_ref, req_prec)) { + + Nvals = Nvals_ref; + rhomin = rhomin_ref; + rhomax = rhomax_ref; + prec = req_prec; + //rho_tbl = Vect_DP (Nvals + 1); + //I_tbl = Vect_DP (Nvals + 1); + rho_tbl = new DP[Nvals + 1]; + I_tbl = new DP[Nvals + 1]; + + logalpha = log(rhomax/rhomin)/Nvals; + alpha = exp(logalpha); + + for (int i = 0; i <= Nvals ; ++i) { + rho_tbl[i] = rhomin * pow(alpha, DP(i)); + //I_tbl[i] = I_integral (rho_tbl[i], req_prec); + I_tbl[i] = function (rho_tbl[i], req_prec); + } + + (*this).Save(); + } + } + + + DP I_table::Return_val (DP req_rho) { + + //cout << "requesting I of " << req_rho << endl; + + DP used_rho = fabs(req_rho); + + if (used_rho < rhomin || used_rho >= rhomax) + { + cerr << "requesting I of " << setprecision(16) << used_rho << " with rho_min = " << rhomin << endl; + //return (I_integral (used_rho, req_prec)); + return (function (used_rho, prec)); + } + + else { + int index = int(log(used_rho/rhomin)/logalpha); // index between 0 and Nvals - 1 + + // Do linear interpolation between values at index and index + 1 + + return((rho_tbl[index + 1] - used_rho) * I_tbl[index] + (used_rho - rho_tbl[index]) * I_tbl[index + 1])/(rho_tbl[index + 1] - rho_tbl[index]); + + } + } + + void I_table::Save () { + + stringstream outfile_strstream; + outfile_strstream << "I_rhomin_" << rhomin << "_rhomax_" << rhomax << "_Nvals_" << Nvals << "_prec_" << prec << ".dat"; + string outfile_str = outfile_strstream.str(); + const char* outfilename = outfile_str.c_str(); + + ofstream outfile; + outfile.open(outfilename); + outfile.precision(16); + + outfile << rho_tbl[0] << "\t" << I_tbl[0]; + for (int i = 1; i <= Nvals; ++i) outfile << endl << rho_tbl[i] << "\t" << I_tbl[i]; + + outfile.close(); + } + + bool I_table::Load (DP rhomin_ref, DP rhomax_ref, int Nvals_ref, DP req_prec) { + + stringstream infile_strstream; + infile_strstream << "I_rhomin_" << rhomin_ref << "_rhomax_" << rhomax_ref << "_Nvals_" << Nvals_ref << "_prec_" << req_prec << ".dat"; + string infile_str = infile_strstream.str(); + const char* infilename = infile_str.c_str(); + + ifstream infile; + infile.open(infilename); + + if (infile.fail()) return (false); + + Nvals = Nvals_ref; + rhomin = rhomin_ref; + rhomax = rhomax_ref; + prec = req_prec; + //rho_tbl = Vect_DP (Nvals + 1); + //I_tbl = Vect_DP (Nvals + 1); + rho_tbl = new DP[Nvals + 1]; + I_tbl = new DP[Nvals + 1]; + + logalpha = log(rhomax/rhomin)/Nvals; + alpha = exp(logalpha); + + int i = 0; + while (infile.peek() != EOF) { + infile >> rho_tbl[i] >> I_tbl[i]; + i++; + } + + if (i != Nvals + 1) JSCerror("Bad input file in I.Load."); + + return(true); + } + + + // ********************* class Integral_table ****************************** + + Integral_table::Integral_table (DP (*function_ref) (DP, DP, int), const char* filenameprefix, + DP rhomin_ref, DP rhomax_ref, int Nvals_ref, DP req_prec, int max_nr_pts) + { + function = function_ref; + + if (!(*this).Load(filenameprefix, rhomin_ref, rhomax_ref, Nvals_ref, req_prec, max_nr_pts)) { + + Nvals = Nvals_ref; + rhomin = rhomin_ref; + rhomax = rhomax_ref; + prec = req_prec; + maxnrpts = max_nr_pts; + + rho_tbl = new DP[Nvals + 1]; + I_tbl = new DP[Nvals + 1]; + + logalpha = log(rhomax/rhomin)/Nvals; + alpha = exp(logalpha); + + for (int i = 0; i <= Nvals ; ++i) { + rho_tbl[i] = rhomin * pow(alpha, DP(i)); + I_tbl[i] = function (rho_tbl[i], req_prec, max_nr_pts); + } + + (*this).Save(filenameprefix); + } + } + + DP Integral_table::Return_val (DP req_rho) { + + //cout << "requesting I of " << req_rho << endl; + + if (req_rho < rhomin || req_rho >= rhomax) + { + cerr << "Requesting I of " << setprecision(16) << req_rho << " with rho_min = " << rhomin << endl; + return (function (req_rho, prec, maxnrpts)); + } + + else { + int index = int(log(req_rho/rhomin)/logalpha); // index between 0 and Nvals - 1 + + // Do linear interpolation between values at index and index + 1 + + return((rho_tbl[index + 1] - req_rho) * I_tbl[index] + (req_rho - rho_tbl[index]) * I_tbl[index + 1])/(rho_tbl[index + 1] - rho_tbl[index]); + } + } + + void Integral_table::Save (const char* filenameprefix) { + + stringstream outfile_strstream; + outfile_strstream << "Integral_table_" << filenameprefix << "_rhomin_" << rhomin << "_rhomax_" << rhomax + << "_Nvals_" << Nvals << "_prec_" << prec << "_maxnrpts_" << maxnrpts << ".dat"; + string outfile_str = outfile_strstream.str(); + const char* outfilename = outfile_str.c_str(); + + ofstream outfile; + outfile.open(outfilename); + outfile.precision(16); + + outfile << rho_tbl[0] << "\t" << I_tbl[0]; + for (int i = 1; i <= Nvals; ++i) outfile << endl << rho_tbl[i] << "\t" << I_tbl[i]; + + outfile.close(); + } + + bool Integral_table::Load (const char* filenameprefix, DP rhomin_ref, DP rhomax_ref, int Nvals_ref, DP req_prec, int max_nr_pts) { + + stringstream infile_strstream; + infile_strstream << "Integral_table_" << filenameprefix << "_rhomin_" << rhomin_ref << "_rhomax_" << rhomax_ref << "_Nvals_" << Nvals_ref << "_prec_" << req_prec << "_maxnrpts_" << max_nr_pts << ".dat"; + string infile_str = infile_strstream.str(); + const char* infilename = infile_str.c_str(); + + ifstream infile; + infile.open(infilename); + + if (infile.fail()) { + //cout << "Failed to load file " << infilename << endl; + return (false); + } + + Nvals = Nvals_ref; + rhomin = rhomin_ref; + rhomax = rhomax_ref; + prec = req_prec; + maxnrpts = max_nr_pts; + + rho_tbl = new DP[Nvals + 1]; + I_tbl = new DP[Nvals + 1]; + + logalpha = log(rhomax/rhomin)/Nvals; + alpha = exp(logalpha); + + int i = 0; + while (infile.peek() != EOF) { + infile >> rho_tbl[i] >> I_tbl[i]; + i++; + } + + if (i != Nvals + 1) JSCerror("Bad input file in Integral_table.Load."); + + return(true); + } + + + + // *********************************** Simple Riemann sum ********************************************* + + DP Integrate_Riemann (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin, DP xmax, int Npts) + { + if (xmax < xmin) return(-Integrate_Riemann (function, args, arg_to_integ, xmax, xmin, Npts)); + + //cout << "Calling Integrate on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + DP dx = (xmax - xmin)/Npts; + + DP result = 0.0; + + for (int i = 0; i < Npts; ++i) { + args[arg_to_integ] = xmin + (i + 0.5) * dx; + result += function (args); + } + result *= dx; + + return(result); + } + + DP Integrate_Riemann_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, int Npts) + { + if (xmax < xmin) return(-Integrate_Riemann_using_table (function, args, arg_to_integ, Itable, xmax, xmin, Npts)); + + //cout << "Calling Integrate on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + DP dx = (xmax - xmin)/Npts; + + DP result = 0.0; + + for (int i = 0; i < Npts; ++i) { + args[arg_to_integ] = xmin + (i + 0.5) * dx; + result += function (args, Itable); + } + result *= dx; + + return(result); + } + + + + // *********************************** Adaptive Riemann sum (straight recursive) ******************************************* + + DP Integrate_rec_main (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin, DP xmax, DP req_prec, int rec_level, int max_rec_level, DP* f) + { + //cout << "Recursion level " << rec_level << endl; + + //cout << "Calling Integrate_rec on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + if (rec_level > max_rec_level) { + //cout << "Warning: integral didn't converge between " << setprecision(10) << xmin << " and " << xmax << endl; + return((xmax - xmin) * (f[0] + f[1] + f[2])/3.0); + } + //if (rec_level > 16) JSCerror("Recursion level too high in Integrate_rec."); + + DP* f1 = new DP[9]; + + DP dx = (xmax - xmin)/9.0; + + args[arg_to_integ] = xmin + 0.5 * dx; + f1[0] = function (args); + f1[1] = f[0]; + args[arg_to_integ] = xmin + 2.5 * dx; + f1[2] = function (args); + args[arg_to_integ] = xmin + 3.5 * dx; + f1[3] = function (args); + f1[4] = f[1]; + args[arg_to_integ] = xmin + 5.5 * dx; + f1[5] = function (args); + args[arg_to_integ] = xmin + 6.5 * dx; + f1[6] = function (args); + f1[7] = f[2]; + args[arg_to_integ] = xmin + 8.5 * dx; + f1[8] = function (args); + + DP I1_pre = 3.0 * dx * f[0]; + DP I1 = dx * (f1[0] + f1[1] + f1[2]); + + if (fabs(I1 - I1_pre) > req_prec || rec_level < 5) + I1 = Integrate_rec_main (function, args, arg_to_integ, xmin, xmin + 3.0 * dx, req_prec, rec_level + 1, max_rec_level, f1); + + DP I2_pre = 3.0 * dx * f[1]; + DP I2 = dx * (f1[3] + f1[4] + f1[5]); + + if (fabs(I2 - I2_pre) > req_prec || rec_level < 5) + I2 = Integrate_rec_main (function, args, arg_to_integ, xmin + 3.0 * dx, xmin + 6.0 * dx, req_prec, rec_level + 1, max_rec_level, &f1[3]); + + DP I3_pre = 3.0 * dx * f[2]; + DP I3 = dx * (f1[6] + f1[7] + f1[8]); + + if (fabs(I3 - I3_pre) > req_prec || rec_level < 5) + I3 = Integrate_rec_main (function, args, arg_to_integ, xmin + 6.0 * dx, xmax, req_prec, rec_level + 1, max_rec_level, &f1[6]); + + delete[] f1; + + return(I1 + I2 + I3); + } + + DP Integrate_rec (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin, DP xmax, DP req_prec, int max_rec_level) + { + if (xmax < xmin) return(-Integrate_rec(function, args, arg_to_integ, xmax, xmin, req_prec, max_rec_level)); + + //cout << "Calling Integrate on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + DP* f = new DP[3]; + DP dx = (xmax - xmin)/3.0; + + DP sum_fabs_f = 0.0; + + for (int i = 0; i < 3; ++i) { + args[arg_to_integ] = xmin + (i + 0.5) * dx; + f[i] = function (args); + sum_fabs_f += fabs(f[i]); + } + + DP req_prec_rec = sum_fabs_f > 1.0 ? sum_fabs_f * req_prec : req_prec; + + //cout << "In Integrate: sum_fabs_f = " << sum_fabs_f << endl; + + DP answer = Integrate_rec_main (function, args, arg_to_integ, xmin, xmax, req_prec_rec, 0, max_rec_level, f); + + delete[] f; + + return(answer); + } + + DP Integrate_rec_main_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int rec_level, int max_rec_level, DP* f) + { + //cout << "Recursion level " << rec_level << endl; + + //cout << "Calling Integrate_rec on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + if (rec_level > max_rec_level) { + //cout << "Warning: integral didn't converge between " << setprecision(10) << xmin << " and " << xmax << endl; + return((xmax - xmin) * (f[0] + f[1] + f[2])/3.0); + } + //if (rec_level > 16) JSCerror("Recursion level too high in Integrate_rec."); + + DP* f1 = new DP[9]; + + DP dx = (xmax - xmin)/9.0; + + args[arg_to_integ] = xmin + 0.5 * dx; + f1[0] = function (args, Itable); + f1[1] = f[0]; + args[arg_to_integ] = xmin + 2.5 * dx; + f1[2] = function (args, Itable); + args[arg_to_integ] = xmin + 3.5 * dx; + f1[3] = function (args, Itable); + f1[4] = f[1]; + args[arg_to_integ] = xmin + 5.5 * dx; + f1[5] = function (args, Itable); + args[arg_to_integ] = xmin + 6.5 * dx; + f1[6] = function (args, Itable); + f1[7] = f[2]; + args[arg_to_integ] = xmin + 8.5 * dx; + f1[8] = function (args, Itable); + + DP I1_pre = 3.0 * dx * f[0]; + DP I1 = dx * (f1[0] + f1[1] + f1[2]); + + if (fabs(I1 - I1_pre) > req_prec || rec_level < 5) + I1 = Integrate_rec_main_using_table (function, args, arg_to_integ, Itable, xmin, xmin + 3.0 * dx, req_prec, rec_level + 1, max_rec_level, f1); + + DP I2_pre = 3.0 * dx * f[1]; + DP I2 = dx * (f1[3] + f1[4] + f1[5]); + + if (fabs(I2 - I2_pre) > req_prec || rec_level < 5) + I2 = Integrate_rec_main_using_table (function, args, arg_to_integ, Itable, xmin + 3.0 * dx, xmin + 6.0 * dx, req_prec, rec_level + 1, max_rec_level, &f1[3]); + + DP I3_pre = 3.0 * dx * f[2]; + DP I3 = dx * (f1[6] + f1[7] + f1[8]); + + if (fabs(I3 - I3_pre) > req_prec || rec_level < 5) + I3 = Integrate_rec_main_using_table (function, args, arg_to_integ, Itable, xmin + 6.0 * dx, xmax, req_prec, rec_level + 1, max_rec_level, &f1[6]); + + delete[] f1; + + return(I1 + I2 + I3); + } + + DP Integrate_rec_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int max_rec_level) + { + if (xmax < xmin) return(-Integrate_rec_using_table (function, args, arg_to_integ, Itable, xmax, xmin, req_prec, max_rec_level)); + + //cout << "Calling Integrate on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + DP* f = new DP[3]; + DP dx = (xmax - xmin)/3.0; + + DP sum_fabs_f = 0.0; + + for (int i = 0; i < 3; ++i) { + args[arg_to_integ] = xmin + (i + 0.5) * dx; + f[i] = function (args, Itable); + sum_fabs_f += fabs(f[i]); + } + + DP req_prec_rec = sum_fabs_f > 1.0 ? sum_fabs_f * req_prec : req_prec; + + //cout << "In Integrate: sum_fabs_f = " << sum_fabs_f << endl; + + DP answer = Integrate_rec_main_using_table (function, args, arg_to_integ, Itable, xmin, xmax, req_prec_rec, 0, max_rec_level, f); + + delete[] f; + + return(answer); + } + + DP Integrate_rec_main_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int rec_level, int max_rec_level, DP* f, ofstream& outfile) + { + //cout << "Recursion level " << rec_level << endl; + + //cout << "Calling Integrate_rec on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + if (rec_level > max_rec_level) { + //cout << "Warning: integral didn't converge between " << setprecision(10) << xmin << " and " << xmax << endl; + return((xmax - xmin) * (f[0] + f[1] + f[2])/3.0); + } + //if (rec_level > 16) JSCerror("Recursion level too high in Integrate_rec."); + + DP* f1 = new DP[9]; + + DP dx = (xmax - xmin)/9.0; + + args[arg_to_integ] = xmin + 0.5 * dx; + f1[0] = function (args, Itable); + //outfile << args[arg_to_integ] << "\t" << f1[0] << endl; + outfile << args << "\t" << f1[0] << endl; + f1[1] = f[0]; + args[arg_to_integ] = xmin + 2.5 * dx; + f1[2] = function (args, Itable); + //outfile << args[arg_to_integ] << "\t" << f1[2] << endl; + outfile << args << "\t" << f1[2] << endl; + args[arg_to_integ] = xmin + 3.5 * dx; + f1[3] = function (args, Itable); + //outfile << args[arg_to_integ] << "\t" << f1[3] << endl; + outfile << args << "\t" << f1[3] << endl; + f1[4] = f[1]; + args[arg_to_integ] = xmin + 5.5 * dx; + f1[5] = function (args, Itable); + //outfile << args[arg_to_integ] << "\t" << f1[5] << endl; + outfile << args << "\t" << f1[5] << endl; + args[arg_to_integ] = xmin + 6.5 * dx; + f1[6] = function (args, Itable); + //outfile << args[arg_to_integ] << "\t" << f1[6] << endl; + outfile << args << "\t" << f1[6] << endl; + f1[7] = f[2]; + args[arg_to_integ] = xmin + 8.5 * dx; + f1[8] = function (args, Itable); + //outfile << args[arg_to_integ] << "\t" << f1[8] << endl; + outfile << args << "\t" << f1[8] << endl; + + DP I1_pre = 3.0 * dx * f[0]; + DP I1 = dx * (f1[0] + f1[1] + f1[2]); + + if (fabs(I1 - I1_pre) > req_prec || rec_level < 5) + I1 = Integrate_rec_main_using_table (function, args, arg_to_integ, Itable, xmin, xmin + 3.0 * dx, req_prec, rec_level + 1, max_rec_level, f1, outfile); + + DP I2_pre = 3.0 * dx * f[1]; + DP I2 = dx * (f1[3] + f1[4] + f1[5]); + + if (fabs(I2 - I2_pre) > req_prec || rec_level < 5) + I2 = Integrate_rec_main_using_table (function, args, arg_to_integ, Itable, xmin + 3.0 * dx, xmin + 6.0 * dx, req_prec, rec_level + 1, max_rec_level, &f1[3], outfile); + + DP I3_pre = 3.0 * dx * f[2]; + DP I3 = dx * (f1[6] + f1[7] + f1[8]); + + if (fabs(I3 - I3_pre) > req_prec || rec_level < 5) + I3 = Integrate_rec_main_using_table (function, args, arg_to_integ, Itable, xmin + 6.0 * dx, xmax, req_prec, rec_level + 1, max_rec_level, &f1[6], outfile); + + delete[] f1; + + return(I1 + I2 + I3); + } + + DP Integrate_rec_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int max_rec_level, ofstream& outfile) + { + if (xmax < xmin) return(-Integrate_rec_using_table (function, args, arg_to_integ, Itable, xmax, xmin, req_prec, max_rec_level, outfile)); + + //cout << "Calling Integrate on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + DP* f = new DP[3]; + DP dx = (xmax - xmin)/3.0; + + DP sum_fabs_f = 0.0; + + for (int i = 0; i < 3; ++i) { + args[arg_to_integ] = xmin + (i + 0.5) * dx; + f[i] = function (args, Itable); + //outfile << args[arg_to_integ] << "\t" << f[i] << endl; + outfile << args << "\t" << f[i] << endl; + sum_fabs_f += fabs(f[i]); + } + + DP req_prec_rec = sum_fabs_f > 1.0 ? sum_fabs_f * req_prec : req_prec; + + //cout << "In Integrate: sum_fabs_f = " << sum_fabs_f << endl; + + DP answer = Integrate_rec_main_using_table (function, args, arg_to_integ, Itable, xmin, xmax, req_prec_rec, 0, max_rec_level, f, outfile); + + delete[] f; + + return(answer); + } + + DP Integrate_rec_main_using_table_and_file (DP (*function) (Vect_DP, I_table, ofstream&), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int rec_level, int max_rec_level, DP* f, ofstream& outfile) + { + //cout << "Recursion level " << rec_level << endl; + + //cout << "Calling Integrate_rec on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + if (rec_level > max_rec_level) { + //cout << "Warning: integral didn't converge between " << setprecision(10) << xmin << " and " << xmax << endl; + return((xmax - xmin) * (f[0] + f[1] + f[2])/3.0); + } + //if (rec_level > 16) JSCerror("Recursion level too high in Integrate_rec."); + + DP* f1 = new DP[9]; + + DP dx = (xmax - xmin)/9.0; + + args[arg_to_integ] = xmin + 0.5 * dx; + f1[0] = function (args, Itable, outfile); + //outfile << args[arg_to_integ] << "\t" << f1[0] << endl; + outfile << args << "\t" << f1[0] << endl; + f1[1] = f[0]; + args[arg_to_integ] = xmin + 2.5 * dx; + f1[2] = function (args, Itable, outfile); + //outfile << args[arg_to_integ] << "\t" << f1[2] << endl; + outfile << args << "\t" << f1[2] << endl; + args[arg_to_integ] = xmin + 3.5 * dx; + f1[3] = function (args, Itable, outfile); + //outfile << args[arg_to_integ] << "\t" << f1[3] << endl; + outfile << args << "\t" << f1[3] << endl; + f1[4] = f[1]; + args[arg_to_integ] = xmin + 5.5 * dx; + f1[5] = function (args, Itable, outfile); + //outfile << args[arg_to_integ] << "\t" << f1[5] << endl; + outfile << args << "\t" << f1[5] << endl; + args[arg_to_integ] = xmin + 6.5 * dx; + f1[6] = function (args, Itable, outfile); + //outfile << args[arg_to_integ] << "\t" << f1[6] << endl; + outfile << args << "\t" << f1[6] << endl; + f1[7] = f[2]; + args[arg_to_integ] = xmin + 8.5 * dx; + f1[8] = function (args, Itable, outfile); + //outfile << args[arg_to_integ] << "\t" << f1[8] << endl; + outfile << args << "\t" << f1[8] << endl; + + DP I1_pre = 3.0 * dx * f[0]; + DP I1 = dx * (f1[0] + f1[1] + f1[2]); + + if (fabs(I1 - I1_pre) > req_prec || rec_level < 5) + I1 = Integrate_rec_main_using_table_and_file (function, args, arg_to_integ, Itable, xmin, xmin + 3.0 * dx, req_prec, rec_level + 1, max_rec_level, f1, outfile); + + DP I2_pre = 3.0 * dx * f[1]; + DP I2 = dx * (f1[3] + f1[4] + f1[5]); + + if (fabs(I2 - I2_pre) > req_prec || rec_level < 5) + I2 = Integrate_rec_main_using_table_and_file (function, args, arg_to_integ, Itable, xmin + 3.0 * dx, xmin + 6.0 * dx, req_prec, rec_level + 1, max_rec_level, &f1[3], outfile); + + DP I3_pre = 3.0 * dx * f[2]; + DP I3 = dx * (f1[6] + f1[7] + f1[8]); + + if (fabs(I3 - I3_pre) > req_prec || rec_level < 5) + I3 = Integrate_rec_main_using_table_and_file (function, args, arg_to_integ, Itable, xmin + 6.0 * dx, xmax, req_prec, rec_level + 1, max_rec_level, &f1[6], outfile); + + delete[] f1; + + return(I1 + I2 + I3); + } + + DP Integrate_rec_using_table_and_file (DP (*function) (Vect_DP, I_table, ofstream&), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int max_rec_level, ofstream& outfile) + { + if (xmax < xmin) return(-Integrate_rec_using_table_and_file (function, args, arg_to_integ, Itable, xmax, xmin, req_prec, max_rec_level, outfile)); + + //cout << "Calling Integrate on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + DP* f = new DP[3]; + DP dx = (xmax - xmin)/3.0; + + DP sum_fabs_f = 0.0; + + for (int i = 0; i < 3; ++i) { + args[arg_to_integ] = xmin + (i + 0.5) * dx; + f[i] = function (args, Itable, outfile); + //outfile << args[arg_to_integ] << "\t" << f[i] << endl; + outfile << args << "\t" << f[i] << endl; + sum_fabs_f += fabs(f[i]); + } + + DP req_prec_rec = sum_fabs_f > 1.0 ? sum_fabs_f * req_prec : req_prec; + + //cout << "In Integrate: sum_fabs_f = " << sum_fabs_f << endl; + + DP answer = Integrate_rec_main_using_table_and_file (function, args, arg_to_integ, Itable, xmin, xmax, req_prec_rec, 0, max_rec_level, f, outfile); + + delete[] f; + + return(answer); + } + + // THESE TWO FUNCTIONS HAVE NOT BEEN TESTED !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! + + DP Integrate_exp_rec_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP eta_old, DP req_prec, int rec_level, int max_rec_level, DP* f, ofstream& outfile) + { + //cout << "Recursion level " << rec_level << endl; + + //cout << "Calling Integrate_rec on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + DP dx_over_x_old = 0.5 * (eta_old - 1.0/eta_old); + if (rec_level > max_rec_level) { + //cout << "Warning: integral didn't converge between " << setprecision(10) << xmin << " and " << xmax << endl; + //return((xmax - xmin) * (f[0] + f[1] + f[2])/3.0); + return(dx_over_x_old * xmin * eta_old * (f[0] + eta_old * (f[1] + eta_old * f[2]))); + } + //if (rec_level > 16) JSCerror("Recursion level too high in Integrate_rec."); + + DP* f1 = new DP[9]; + DP* x_pts = new DP[9]; + DP eta = pow(xmax/xmin, 1.0/18.0); + DP eta_sq = eta * eta; + DP dx_over_x = 0.5 * (eta - 1.0/eta); + + x_pts[0] = xmin * eta; + for (int i = 1; i < 9; ++i) x_pts[i] = x_pts[i-1] * eta_sq; + + args[arg_to_integ] = x_pts[0]; + f1[0] = function (args, Itable); + outfile << args[arg_to_integ] << "\t" << f1[0] << endl; + f1[1] = f[0]; + args[arg_to_integ] = x_pts[2]; + f1[2] = function (args, Itable); + outfile << args[arg_to_integ] << "\t" << f1[2] << endl; + args[arg_to_integ] = x_pts[3]; + f1[3] = function (args, Itable); + outfile << args[arg_to_integ] << "\t" << f1[3] << endl; + f1[4] = f[1]; + args[arg_to_integ] = x_pts[5]; + f1[5] = function (args, Itable); + outfile << args[arg_to_integ] << "\t" << f1[5] << endl; + args[arg_to_integ] = x_pts[6]; + f1[6] = function (args, Itable); + outfile << args[arg_to_integ] << "\t" << f1[6] << endl; + f1[7] = f[2]; + args[arg_to_integ] = x_pts[8]; + f1[8] = function (args, Itable); + outfile << args[arg_to_integ] << "\t" << f1[8] << endl; + + DP I1_pre = dx_over_x_old * x_pts[1] * f1[1]; + DP I1 = dx_over_x * (x_pts[0] * f1[0] + x_pts[1] * f1[1] + x_pts[2] * f1[2]); + + if (fabs(I1 - I1_pre) > req_prec || rec_level < 5) + I1 = Integrate_exp_rec_using_table (function, args, arg_to_integ, Itable, xmin, x_pts[2] * eta, eta, req_prec, rec_level + 1, max_rec_level, f1, outfile); + + DP I2_pre = dx_over_x_old * x_pts[4] * f1[4]; + DP I2 = dx_over_x * (x_pts[3] * f1[3] + x_pts[4] * f1[4] + x_pts[5] * f1[5]); + + if (fabs(I2 - I2_pre) > req_prec || rec_level < 5) + I2 = Integrate_exp_rec_using_table (function, args, arg_to_integ, Itable, x_pts[2] * eta, x_pts[5] * eta, eta, req_prec, rec_level + 1, max_rec_level, &f1[3], outfile); + + DP I3_pre = dx_over_x_old * f1[7]; + DP I3 = dx_over_x * (x_pts[6] * f1[6] + x_pts[7] * f1[7] + x_pts[8] * f1[8]); + + if (fabs(I3 - I3_pre) > req_prec || rec_level < 5) + I3 = Integrate_exp_rec_using_table (function, args, arg_to_integ, Itable, x_pts[5] * eta, xmax, eta, req_prec, rec_level + 1, max_rec_level, &f1[6], outfile); + + delete[] f1; + + return(I1 + I2 + I3); + } + + DP Integrate_exp_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, + DP xmin, DP xmax, DP req_prec, int max_rec_level, ofstream& outfile) + { + // This uses an exponential progression of sampling points. + + if (xmin <= 0.0) JSCerror("Use xmin > 0.0 in Integrate_exp."); + if (xmax <= 0.0) JSCerror("Use xmax > 0.0 in Integrate_exp."); + + if (xmax < xmin) return(-Integrate_exp_using_table (function, args, arg_to_integ, Itable, xmax, xmin, req_prec, max_rec_level, outfile)); + + //cout << "Calling Integrate on argument " << arg_to_integ << " between " << xmin << " and " << xmax << " to prec " << req_prec << endl; + + DP* f = new DP[3]; + DP eta = pow(xmax/xmin, 1.0/6.0); + //DP dx_over_x = 0.5 * (eta - 1.0/eta); + + DP sum_fabs_f = 0.0; + + args[arg_to_integ] = xmin * eta; + f[0] = function (args, Itable); + outfile << args[arg_to_integ] << "\t" << f[0] << endl; + sum_fabs_f += fabs(f[0]); + args[arg_to_integ] *= eta * eta; + f[1] = function (args, Itable); + outfile << args[arg_to_integ] << "\t" << f[1] << endl; + sum_fabs_f += fabs(f[1]); + args[arg_to_integ] *= eta * eta; + f[2] = function (args, Itable); + outfile << args[arg_to_integ] << "\t" << f[2] << endl; + sum_fabs_f += fabs(f[2]); + + DP req_prec_rec = sum_fabs_f > 1.0 ? sum_fabs_f * req_prec : req_prec; + + //cout << "In Integrate: sum_fabs_f = " << sum_fabs_f << endl; + + DP answer = Integrate_exp_rec_using_table (function, args, arg_to_integ, Itable, xmin, xmax, eta, req_prec_rec, 0, max_rec_level, f, outfile); + + delete[] f; + + return(answer); + } + + + + // *********************************** Adaptive Riemann sum (better implementation) ******************************************* + + std::ostream& operator<< (std::ostream& s, const Integral_result& res) + { + s << res.integ_est << "\t" << res.abs_prec << "\t" << res.rel_prec << "\t" << res.n_vals; + + return(s); + } + + Integral_data::~Integral_data() + { + if (data != 0) delete[] data; + if (abs_d2f_dx != 0) delete[] abs_d2f_dx; + } + + Integral_data::Integral_data (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin_ref, DP xmax_ref) + { + xmin = xmin_ref; + xmax = xmax_ref; + integ_res.n_vals = 9; + + data = new data_pt[integ_res.n_vals]; + abs_d2f_dx = new DP[integ_res.n_vals/3]; + + integ_res.integ_est = 0.0; + + // Initialize x, f and dx: + DP dx = (xmax - xmin)/9.0; + for (int i = 0; i < 9; ++i) { + data[i].x = xmin + (i + 0.5) * dx; + args[arg_to_integ] = data[i].x; + data[i].f = function (args); + data[i].dx = dx; + integ_res.integ_est += data[i].dx * data[i].f; + } + + max_abs_d2f_dx = 0.0; + integ_res.abs_prec = 0.0; + for (int j = 0; j < 3; ++j) { + abs_d2f_dx[j] = dx * fabs(data[3*j].f - 2.0 * data[3*j + 1].f + data[3*j + 2].f); + max_abs_d2f_dx = JSC::max(max_abs_d2f_dx, abs_d2f_dx[j]); + integ_res.abs_prec += abs_d2f_dx[j]; + } + + integ_res.rel_prec = integ_res.abs_prec/integ_res.integ_est; + + //for (int in = 0; in < n_vals; ++in) + //cout << in << "\t" << data[in].x << "\t" << data[in].f << "\t" << data[in].dx << endl; + } + + Integral_data::Integral_data (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, DP xmin_ref, DP xmax_ref) + { + xmin = xmin_ref; + xmax = xmax_ref; + integ_res.n_vals = 9; + + data = new data_pt[integ_res.n_vals]; + abs_d2f_dx = new DP[integ_res.n_vals/3]; + + integ_res.integ_est = 0.0; + + // Initialize x, f and dx: + DP dx = (xmax - xmin)/9.0; + for (int i = 0; i < 9; ++i) { + data[i].x = xmin + (i + 0.5) * dx; + args[arg_to_integ] = data[i].x; + data[i].f = function (args, Itable); + data[i].dx = dx; + integ_res.integ_est += data[i].dx * data[i].f; + } + + max_abs_d2f_dx = 0.0; + integ_res.abs_prec = 0.0; + for (int j = 0; j < 3; ++j) { + abs_d2f_dx[j] = dx * fabs(data[3*j].f - 2.0 * data[3*j + 1].f + data[3*j + 2].f); + max_abs_d2f_dx = JSC::max(max_abs_d2f_dx, abs_d2f_dx[j]); + integ_res.abs_prec += abs_d2f_dx[j]; + } + + integ_res.rel_prec = integ_res.abs_prec/integ_res.integ_est; + + //for (int in = 0; in < n_vals; ++in) + //cout << in << "\t" << data[in].x << "\t" << data[in].f << "\t" << data[in].dx << endl; + } + + Integral_data::Integral_data (DP (*function) (Vect_DP, Integral_table), Vect_DP& args, int arg_to_integ, Integral_table Itable, DP xmin_ref, DP xmax_ref) + { + xmin = xmin_ref; + xmax = xmax_ref; + integ_res.n_vals = 9; + + data = new data_pt[integ_res.n_vals]; + abs_d2f_dx = new DP[integ_res.n_vals/3]; + + integ_res.integ_est = 0.0; + + // Initialize x, f and dx: + DP dx = (xmax - xmin)/9.0; + for (int i = 0; i < 9; ++i) { + data[i].x = xmin + (i + 0.5) * dx; + args[arg_to_integ] = data[i].x; + data[i].f = function (args, Itable); + data[i].dx = dx; + integ_res.integ_est += data[i].dx * data[i].f; + } + + max_abs_d2f_dx = 0.0; + integ_res.abs_prec = 0.0; + for (int j = 0; j < 3; ++j) { + abs_d2f_dx[j] = dx * fabs(data[3*j].f - 2.0 * data[3*j + 1].f + data[3*j + 2].f); + max_abs_d2f_dx = JSC::max(max_abs_d2f_dx, abs_d2f_dx[j]); + integ_res.abs_prec += abs_d2f_dx[j]; + } + + integ_res.rel_prec = integ_res.abs_prec/integ_res.integ_est; + + //for (int in = 0; in < n_vals; ++in) + //cout << in << "\t" << data[in].x << "\t" << data[in].f << "\t" << data[in].dx << endl; + } + + void Integral_data::Save (ofstream& outfile) + { + // Reset file writing position + outfile.seekp(0); + int i = 0; + while (i < integ_res.n_vals && !(data[i].x == 0.0 && data[i].f == 0.0)) { + outfile << setprecision(16) << data[i].x << "\t" << data[i].f << endl; + i++; + } + } + + void Integral_data::Improve_estimate (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, int max_nr_pts) + { + // We generate a max of 3* n_vals points + data_pt* new_data = new data_pt[3* integ_res.n_vals]; + DP* new_abs_d2f_dx = new DP[integ_res.n_vals]; + + // Check points in batches of 3; if needed, improve + int threei = 0; + int index_new = 0; + int i, j; + + integ_res.abs_prec = 0.0; + integ_res.integ_est = 0.0; + + DP new_max_abs_d2f_dx = 0.0; + + for (i = 0; i < integ_res.n_vals/3; ++i) { + + threei = 3 * i; + + if (abs_d2f_dx[i] <= 0.1 * max_abs_d2f_dx || index_new + integ_res.n_vals - threei > max_nr_pts) { + + // simply transfer the data points into new_data + new_abs_d2f_dx[index_new/3] = abs_d2f_dx[i]; + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + for (j = 0; j < 3; ++j) { + new_data[index_new].x = data[threei + j].x; + new_data[index_new].f = data[threei + j].f; + new_data[index_new].dx = data[threei + j].dx; + integ_res.integ_est += new_data[index_new].dx * new_data[index_new].f; + index_new++; + } + integ_res.abs_prec += abs_d2f_dx[i]; + + } + + else { // create six new entries and transfer the three existing ones + + new_data[index_new].dx = data[threei].dx/3.0; + for (j = 1; j < 9; ++j) new_data[index_new + j].dx = new_data[index_new].dx; + + new_data[index_new].x = data[threei].x - new_data[index_new].dx; + new_data[index_new + 1].x = data[threei].x; + new_data[index_new + 2].x = data[threei].x + new_data[index_new].dx; + + new_data[index_new + 3].x = data[threei + 1].x - new_data[index_new].dx; + new_data[index_new + 4].x = data[threei + 1].x; + new_data[index_new + 5].x = data[threei + 1].x + new_data[index_new].dx; + + new_data[index_new + 6].x = data[threei + 2].x - new_data[index_new].dx; + new_data[index_new + 7].x = data[threei + 2].x; + new_data[index_new + 8].x = data[threei + 2].x + new_data[index_new].dx; + + args[arg_to_integ] = new_data[index_new].x; + new_data[index_new].f = function(args); + new_data[index_new + 1].f = data[threei].f; + args[arg_to_integ] = new_data[index_new + 2].x; + new_data[index_new + 2].f = function(args); + + args[arg_to_integ] = new_data[index_new + 3].x; + new_data[index_new + 3].f = function(args); + new_data[index_new + 4].f = data[threei + 1].f; + args[arg_to_integ] = new_data[index_new + 5].x; + new_data[index_new + 5].f = function(args); + + args[arg_to_integ] = new_data[index_new + 6].x; + new_data[index_new + 6].f = function(args); + new_data[index_new + 7].f = data[threei + 2].f; + args[arg_to_integ] = new_data[index_new + 8].x; + new_data[index_new + 8].f = function(args); + + new_abs_d2f_dx[index_new/3] = fabs(new_data[index_new].dx * (new_data[index_new].f - 2.0 * new_data[index_new + 1].f + new_data[index_new + 2].f)); + new_abs_d2f_dx[index_new/3 + 1] = fabs(new_data[index_new].dx * (new_data[index_new + 3].f - 2.0 * new_data[index_new + 4].f + new_data[index_new + 5].f)); + new_abs_d2f_dx[index_new/3 + 2] = fabs(new_data[index_new].dx * (new_data[index_new + 6].f - 2.0 * new_data[index_new + 7].f + new_data[index_new + 8].f)); + + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 1]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 2]); + + integ_res.integ_est += new_data[index_new].dx * (new_data[index_new].f + new_data[index_new + 1].f + new_data[index_new + 2].f + + new_data[index_new + 3].f + new_data[index_new + 4].f + new_data[index_new + 5].f + + new_data[index_new + 6].f + new_data[index_new + 7].f + new_data[index_new + 8].f); + + integ_res.abs_prec += new_abs_d2f_dx[index_new/3] + new_abs_d2f_dx[index_new/3 + 1] + new_abs_d2f_dx[index_new/3 + 2]; + + index_new += 9; + + } // else + + } // for (i < nvals/3) + + integ_res.rel_prec = integ_res.abs_prec/integ_res.integ_est; + + integ_res.n_vals = index_new; + + delete[] data; + data = new_data; + + max_abs_d2f_dx = new_max_abs_d2f_dx; + + delete[] abs_d2f_dx; + abs_d2f_dx = new_abs_d2f_dx; + + return; + } + + void Integral_data::Improve_estimate (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, int max_nr_pts) + { + // We generate a max of 3* n_vals points + data_pt* new_data = new data_pt[3* integ_res.n_vals]; + DP* new_abs_d2f_dx = new DP[integ_res.n_vals]; + + // Check points in batches of 3; if needed, improve + int threei = 0; + int index_new = 0; + int i, j; + + integ_res.abs_prec = 0.0; + integ_res.integ_est = 0.0; + + DP new_max_abs_d2f_dx = 0.0; + + for (i = 0; i < integ_res.n_vals/3; ++i) { + + threei = 3 * i; + + if (abs_d2f_dx[i] <= 0.1 * max_abs_d2f_dx || index_new + integ_res.n_vals - threei > max_nr_pts) { + + // simply transfer the data points into new_data + new_abs_d2f_dx[index_new/3] = abs_d2f_dx[i]; + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + for (j = 0; j < 3; ++j) { + new_data[index_new].x = data[threei + j].x; + new_data[index_new].f = data[threei + j].f; + new_data[index_new].dx = data[threei + j].dx; + integ_res.integ_est += new_data[index_new].dx * new_data[index_new].f; + index_new++; + } + integ_res.abs_prec += abs_d2f_dx[i]; + + } + + else { // create six new entries and transfer the three existing ones + + new_data[index_new].dx = data[threei].dx/3.0; + for (j = 1; j < 9; ++j) new_data[index_new + j].dx = new_data[index_new].dx; + + new_data[index_new].x = data[threei].x - new_data[index_new].dx; + new_data[index_new + 1].x = data[threei].x; + new_data[index_new + 2].x = data[threei].x + new_data[index_new].dx; + + new_data[index_new + 3].x = data[threei + 1].x - new_data[index_new].dx; + new_data[index_new + 4].x = data[threei + 1].x; + new_data[index_new + 5].x = data[threei + 1].x + new_data[index_new].dx; + + new_data[index_new + 6].x = data[threei + 2].x - new_data[index_new].dx; + new_data[index_new + 7].x = data[threei + 2].x; + new_data[index_new + 8].x = data[threei + 2].x + new_data[index_new].dx; + + args[arg_to_integ] = new_data[index_new].x; + new_data[index_new].f = function(args, Itable); + new_data[index_new + 1].f = data[threei].f; + args[arg_to_integ] = new_data[index_new + 2].x; + new_data[index_new + 2].f = function(args, Itable); + + args[arg_to_integ] = new_data[index_new + 3].x; + new_data[index_new + 3].f = function(args, Itable); + new_data[index_new + 4].f = data[threei + 1].f; + args[arg_to_integ] = new_data[index_new + 5].x; + new_data[index_new + 5].f = function(args, Itable); + + args[arg_to_integ] = new_data[index_new + 6].x; + new_data[index_new + 6].f = function(args, Itable); + new_data[index_new + 7].f = data[threei + 2].f; + args[arg_to_integ] = new_data[index_new + 8].x; + new_data[index_new + 8].f = function(args, Itable); + + new_abs_d2f_dx[index_new/3] = fabs(new_data[index_new].dx * (new_data[index_new].f - 2.0 * new_data[index_new + 1].f + new_data[index_new + 2].f)); + new_abs_d2f_dx[index_new/3 + 1] = fabs(new_data[index_new].dx * (new_data[index_new + 3].f - 2.0 * new_data[index_new + 4].f + new_data[index_new + 5].f)); + new_abs_d2f_dx[index_new/3 + 2] = fabs(new_data[index_new].dx * (new_data[index_new + 6].f - 2.0 * new_data[index_new + 7].f + new_data[index_new + 8].f)); + + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 1]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 2]); + + integ_res.integ_est += new_data[index_new].dx * (new_data[index_new].f + new_data[index_new + 1].f + new_data[index_new + 2].f + + new_data[index_new + 3].f + new_data[index_new + 4].f + new_data[index_new + 5].f + + new_data[index_new + 6].f + new_data[index_new + 7].f + new_data[index_new + 8].f); + + integ_res.abs_prec += new_abs_d2f_dx[index_new/3] + new_abs_d2f_dx[index_new/3 + 1] + new_abs_d2f_dx[index_new/3 + 2]; + + index_new += 9; + + } // else + + } // for (i < nvals/3) + + integ_res.rel_prec = integ_res.abs_prec/integ_res.integ_est; + + integ_res.n_vals = index_new; + + delete[] data; + data = new_data; + + max_abs_d2f_dx = new_max_abs_d2f_dx; + + delete[] abs_d2f_dx; + abs_d2f_dx = new_abs_d2f_dx; + + return; + } + + void Integral_data::Improve_estimate (DP (*function) (Vect_DP, Integral_table), Vect_DP& args, int arg_to_integ, Integral_table Itable, int max_nr_pts) + { + // We generate a max of 3* n_vals points + data_pt* new_data = new data_pt[3* integ_res.n_vals]; + DP* new_abs_d2f_dx = new DP[integ_res.n_vals]; + + // Check points in batches of 3; if needed, improve + int threei = 0; + int index_new = 0; + int i, j; + + integ_res.abs_prec = 0.0; + integ_res.integ_est = 0.0; + + DP new_max_abs_d2f_dx = 0.0; + + for (i = 0; i < integ_res.n_vals/3; ++i) { + + threei = 3 * i; + + if (abs_d2f_dx[i] <= 0.1 * max_abs_d2f_dx || index_new + integ_res.n_vals - threei > max_nr_pts) { + + // simply transfer the data points into new_data + new_abs_d2f_dx[index_new/3] = abs_d2f_dx[i]; + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + for (j = 0; j < 3; ++j) { + new_data[index_new].x = data[threei + j].x; + new_data[index_new].f = data[threei + j].f; + new_data[index_new].dx = data[threei + j].dx; + integ_res.integ_est += new_data[index_new].dx * new_data[index_new].f; + index_new++; + } + integ_res.abs_prec += abs_d2f_dx[i]; + + } + + else { // create six new entries and transfer the three existing ones + + new_data[index_new].dx = data[threei].dx/3.0; + for (j = 1; j < 9; ++j) new_data[index_new + j].dx = new_data[index_new].dx; + + new_data[index_new].x = data[threei].x - new_data[index_new].dx; + new_data[index_new + 1].x = data[threei].x; + new_data[index_new + 2].x = data[threei].x + new_data[index_new].dx; + + new_data[index_new + 3].x = data[threei + 1].x - new_data[index_new].dx; + new_data[index_new + 4].x = data[threei + 1].x; + new_data[index_new + 5].x = data[threei + 1].x + new_data[index_new].dx; + + new_data[index_new + 6].x = data[threei + 2].x - new_data[index_new].dx; + new_data[index_new + 7].x = data[threei + 2].x; + new_data[index_new + 8].x = data[threei + 2].x + new_data[index_new].dx; + + args[arg_to_integ] = new_data[index_new].x; + new_data[index_new].f = function(args, Itable); + new_data[index_new + 1].f = data[threei].f; + args[arg_to_integ] = new_data[index_new + 2].x; + new_data[index_new + 2].f = function(args, Itable); + + args[arg_to_integ] = new_data[index_new + 3].x; + new_data[index_new + 3].f = function(args, Itable); + new_data[index_new + 4].f = data[threei + 1].f; + args[arg_to_integ] = new_data[index_new + 5].x; + new_data[index_new + 5].f = function(args, Itable); + + args[arg_to_integ] = new_data[index_new + 6].x; + new_data[index_new + 6].f = function(args, Itable); + new_data[index_new + 7].f = data[threei + 2].f; + args[arg_to_integ] = new_data[index_new + 8].x; + new_data[index_new + 8].f = function(args, Itable); + + new_abs_d2f_dx[index_new/3] = fabs(new_data[index_new].dx * (new_data[index_new].f - 2.0 * new_data[index_new + 1].f + new_data[index_new + 2].f)); + new_abs_d2f_dx[index_new/3 + 1] = fabs(new_data[index_new].dx * (new_data[index_new + 3].f - 2.0 * new_data[index_new + 4].f + new_data[index_new + 5].f)); + new_abs_d2f_dx[index_new/3 + 2] = fabs(new_data[index_new].dx * (new_data[index_new + 6].f - 2.0 * new_data[index_new + 7].f + new_data[index_new + 8].f)); + + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 1]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 2]); + + integ_res.integ_est += new_data[index_new].dx * (new_data[index_new].f + new_data[index_new + 1].f + new_data[index_new + 2].f + + new_data[index_new + 3].f + new_data[index_new + 4].f + new_data[index_new + 5].f + + new_data[index_new + 6].f + new_data[index_new + 7].f + new_data[index_new + 8].f); + + integ_res.abs_prec += new_abs_d2f_dx[index_new/3] + new_abs_d2f_dx[index_new/3 + 1] + new_abs_d2f_dx[index_new/3 + 2]; + + index_new += 9; + + } // else + + } // for (i < nvals/3) + + integ_res.rel_prec = integ_res.abs_prec/integ_res.integ_est; + + integ_res.n_vals = index_new; + + delete[] data; + data = new_data; + + max_abs_d2f_dx = new_max_abs_d2f_dx; + + delete[] abs_d2f_dx; + abs_d2f_dx = new_abs_d2f_dx; + + return; + } + + Integral_result Integrate_optimal (DP (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts) + { + if (xmax < xmin) JSCerror("Use xmax > xmin in Integrate."); + + Integral_data integ_dat (function, args, arg_to_integ, xmin, xmax); + + while ((integ_dat.integ_res.rel_prec > req_rel_prec || integ_dat.integ_res.abs_prec > req_abs_prec) && integ_dat.integ_res.n_vals < max_nr_pts) { + + integ_dat.Improve_estimate (function, args, arg_to_integ, max_nr_pts); + } + + // REMOVE next four line + //ofstream outfile; + //outfile.open("testoptimal.dat"); + //outfile.precision(16); + //integ_dat.Save(outfile); + + return(integ_dat.integ_res); + } + + Integral_result Integrate_optimal_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, + I_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts) + { + if (xmax < xmin) JSCerror("Use xmax > xmin in Integrate."); + + Integral_data integ_dat (function, args, arg_to_integ, Itable, xmin, xmax); + + while ((integ_dat.integ_res.rel_prec > req_rel_prec || integ_dat.integ_res.abs_prec > req_abs_prec) && integ_dat.integ_res.n_vals < max_nr_pts) { + + integ_dat.Improve_estimate (function, args, arg_to_integ, Itable, max_nr_pts); + } + + return(integ_dat.integ_res); + } + + Integral_result Integrate_optimal_using_table (DP (*function) (Vect_DP, Integral_table), Vect_DP& args, int arg_to_integ, + Integral_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts) + { + if (xmax < xmin) JSCerror("Use xmax > xmin in Integrate."); + + Integral_data integ_dat (function, args, arg_to_integ, Itable, xmin, xmax); + + while ((integ_dat.integ_res.rel_prec > req_rel_prec || integ_dat.integ_res.abs_prec > req_abs_prec) && integ_dat.integ_res.n_vals < max_nr_pts) { + + integ_dat.Improve_estimate (function, args, arg_to_integ, Itable, max_nr_pts); + } + + return(integ_dat.integ_res); + } + + Integral_result Integrate_optimal_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, + I_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts, ofstream& outfile) + { + if (xmax < xmin) JSCerror("Use xmax > xmin in Integrate."); + + Integral_data integ_dat (function, args, arg_to_integ, Itable, xmin, xmax); + + while ((integ_dat.integ_res.rel_prec > req_rel_prec || integ_dat.integ_res.abs_prec > req_abs_prec) && integ_dat.integ_res.n_vals < max_nr_pts) { + + integ_dat.Improve_estimate (function, args, arg_to_integ, Itable, max_nr_pts); + + integ_dat.Save (outfile); + + } + + return(integ_dat.integ_res); + } + + + // FOR COMPLEX FUNCTIONS: + // *********************************** Adaptive Riemann sum (better implementation) ******************************************* + + std::ostream& operator<< (std::ostream& s, const Integral_result_CX& res) + { + s << real(res.integ_est) << "\t" << imag(res.integ_est) << "\t" << res.abs_prec << "\t" << res.rel_prec << "\t" << res.n_vals; + + return(s); + } + + Integral_data_CX::~Integral_data_CX() + { + if (data != 0) delete[] data; + if (abs_d2f_dx != 0) delete[] abs_d2f_dx; + } + + Integral_data_CX::Integral_data_CX (complex (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin_ref, DP xmax_ref) + { + xmin = xmin_ref; + xmax = xmax_ref; + integ_res.n_vals = 9; + + data = new data_pt_CX[integ_res.n_vals]; + abs_d2f_dx = new DP[integ_res.n_vals/3]; + + integ_res.integ_est = 0.0; + + // Initialize x, f and dx: + DP dx = (xmax - xmin)/9.0; + for (int i = 0; i < 9; ++i) { + data[i].x = xmin + (i + 0.5) * dx; + args[arg_to_integ] = data[i].x; + data[i].f = function (args); + data[i].dx = dx; + integ_res.integ_est += data[i].dx * data[i].f; + } + + max_abs_d2f_dx = 0.0; + integ_res.abs_prec = 0.0; + for (int j = 0; j < 3; ++j) { + abs_d2f_dx[j] = dx * abs(data[3*j].f - 2.0 * data[3*j + 1].f + data[3*j + 2].f); + max_abs_d2f_dx = JSC::max(max_abs_d2f_dx, abs_d2f_dx[j]); + integ_res.abs_prec += abs_d2f_dx[j]; + } + + integ_res.rel_prec = integ_res.abs_prec/abs(integ_res.integ_est); + + //for (int in = 0; in < n_vals; ++in) + //cout << in << "\t" << data[in].x << "\t" << data[in].f << "\t" << data[in].dx << endl; + } + + // No implementation with I_table yet for complex... + /* + Integral_data::Integral_data (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, DP xmin_ref, DP xmax_ref) + { + xmin = xmin_ref; + xmax = xmax_ref; + integ_res.n_vals = 9; + + data = new data_pt[integ_res.n_vals]; + abs_d2f_dx = new DP[integ_res.n_vals/3]; + + integ_res.integ_est = 0.0; + + // Initialize x, f and dx: + DP dx = (xmax - xmin)/9.0; + for (int i = 0; i < 9; ++i) { + data[i].x = xmin + (i + 0.5) * dx; + args[arg_to_integ] = data[i].x; + data[i].f = function (args, Itable); + data[i].dx = dx; + integ_res.integ_est += data[i].dx * data[i].f; + } + + max_abs_d2f_dx = 0.0; + integ_res.abs_prec = 0.0; + for (int j = 0; j < 3; ++j) { + abs_d2f_dx[j] = dx * fabs(data[3*j].f - 2.0 * data[3*j + 1].f + data[3*j + 2].f); + max_abs_d2f_dx = JSC::max(max_abs_d2f_dx, abs_d2f_dx[j]); + integ_res.abs_prec += abs_d2f_dx[j]; + } + + integ_res.rel_prec = integ_res.abs_prec/integ_res.integ_est; + + //for (int in = 0; in < n_vals; ++in) + //cout << in << "\t" << data[in].x << "\t" << data[in].f << "\t" << data[in].dx << endl; + } + */ + void Integral_data_CX::Save (ofstream& outfile) + { + // Reset file writing position + outfile.seekp(0); + int i = 0; + while (i < integ_res.n_vals && !(data[i].x == 0.0 && data[i].f == 0.0)) { + outfile << setprecision(16) << data[i].x << "\t" << data[i].f << endl; + i++; + } + } + + void Integral_data_CX::Improve_estimate (complex (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, int max_nr_pts) + { + // We generate a max of 3* n_vals points + data_pt_CX* new_data = new data_pt_CX[3* integ_res.n_vals]; + DP* new_abs_d2f_dx = new DP[integ_res.n_vals]; + + // Check points in batches of 3; if needed, improve + int threei = 0; + int index_new = 0; + int i, j; + + integ_res.abs_prec = 0.0; + integ_res.integ_est = 0.0; + + DP new_max_abs_d2f_dx = 0.0; + + for (i = 0; i < integ_res.n_vals/3; ++i) { + + threei = 3 * i; + + if (abs_d2f_dx[i] <= 0.1 * max_abs_d2f_dx || index_new + integ_res.n_vals - threei > max_nr_pts) { + + // simply transfer the data points into new_data + new_abs_d2f_dx[index_new/3] = abs_d2f_dx[i]; + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + for (j = 0; j < 3; ++j) { + new_data[index_new].x = data[threei + j].x; + new_data[index_new].f = data[threei + j].f; + new_data[index_new].dx = data[threei + j].dx; + integ_res.integ_est += new_data[index_new].dx * new_data[index_new].f; + index_new++; + } + integ_res.abs_prec += abs_d2f_dx[i]; + + } + + else { // create six new entries and transfer the three existing ones + + new_data[index_new].dx = data[threei].dx/3.0; + for (j = 1; j < 9; ++j) new_data[index_new + j].dx = new_data[index_new].dx; + + new_data[index_new].x = data[threei].x - new_data[index_new].dx; + new_data[index_new + 1].x = data[threei].x; + new_data[index_new + 2].x = data[threei].x + new_data[index_new].dx; + + new_data[index_new + 3].x = data[threei + 1].x - new_data[index_new].dx; + new_data[index_new + 4].x = data[threei + 1].x; + new_data[index_new + 5].x = data[threei + 1].x + new_data[index_new].dx; + + new_data[index_new + 6].x = data[threei + 2].x - new_data[index_new].dx; + new_data[index_new + 7].x = data[threei + 2].x; + new_data[index_new + 8].x = data[threei + 2].x + new_data[index_new].dx; + + args[arg_to_integ] = new_data[index_new].x; + new_data[index_new].f = function(args); + new_data[index_new + 1].f = data[threei].f; + args[arg_to_integ] = new_data[index_new + 2].x; + new_data[index_new + 2].f = function(args); + + args[arg_to_integ] = new_data[index_new + 3].x; + new_data[index_new + 3].f = function(args); + new_data[index_new + 4].f = data[threei + 1].f; + args[arg_to_integ] = new_data[index_new + 5].x; + new_data[index_new + 5].f = function(args); + + args[arg_to_integ] = new_data[index_new + 6].x; + new_data[index_new + 6].f = function(args); + new_data[index_new + 7].f = data[threei + 2].f; + args[arg_to_integ] = new_data[index_new + 8].x; + new_data[index_new + 8].f = function(args); + + new_abs_d2f_dx[index_new/3] = abs(new_data[index_new].dx * (new_data[index_new].f - 2.0 * new_data[index_new + 1].f + new_data[index_new + 2].f)); + new_abs_d2f_dx[index_new/3 + 1] = abs(new_data[index_new].dx * (new_data[index_new + 3].f - 2.0 * new_data[index_new + 4].f + new_data[index_new + 5].f)); + new_abs_d2f_dx[index_new/3 + 2] = abs(new_data[index_new].dx * (new_data[index_new + 6].f - 2.0 * new_data[index_new + 7].f + new_data[index_new + 8].f)); + + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 1]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 2]); + + integ_res.integ_est += new_data[index_new].dx * (new_data[index_new].f + new_data[index_new + 1].f + new_data[index_new + 2].f + + new_data[index_new + 3].f + new_data[index_new + 4].f + new_data[index_new + 5].f + + new_data[index_new + 6].f + new_data[index_new + 7].f + new_data[index_new + 8].f); + + integ_res.abs_prec += new_abs_d2f_dx[index_new/3] + new_abs_d2f_dx[index_new/3 + 1] + new_abs_d2f_dx[index_new/3 + 2]; + + index_new += 9; + + } // else + + } // for (i < nvals/3) + + integ_res.rel_prec = integ_res.abs_prec/abs(integ_res.integ_est); + + integ_res.n_vals = index_new; + + delete[] data; + data = new_data; + + max_abs_d2f_dx = new_max_abs_d2f_dx; + + delete[] abs_d2f_dx; + abs_d2f_dx = new_abs_d2f_dx; + + return; + } + + // No implementation with I_table yet for complex... + /* + void Integral_data::Improve_estimate (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, int max_nr_pts) + { + // We generate a max of 3* n_vals points + data_pt* new_data = new data_pt[3* integ_res.n_vals]; + DP* new_abs_d2f_dx = new DP[integ_res.n_vals]; + + // Check points in batches of 3; if needed, improve + int threei = 0; + int index_new = 0; + int i, j; + + integ_res.abs_prec = 0.0; + integ_res.integ_est = 0.0; + + DP new_max_abs_d2f_dx = 0.0; + + for (i = 0; i < integ_res.n_vals/3; ++i) { + + threei = 3 * i; + + if (abs_d2f_dx[i] <= 0.1 * max_abs_d2f_dx || index_new + integ_res.n_vals - threei > max_nr_pts) { + + // simply transfer the data points into new_data + new_abs_d2f_dx[index_new/3] = abs_d2f_dx[i]; + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + for (j = 0; j < 3; ++j) { + new_data[index_new].x = data[threei + j].x; + new_data[index_new].f = data[threei + j].f; + new_data[index_new].dx = data[threei + j].dx; + integ_res.integ_est += new_data[index_new].dx * new_data[index_new].f; + index_new++; + } + integ_res.abs_prec += abs_d2f_dx[i]; + + } + + else { // create six new entries and transfer the three existing ones + + new_data[index_new].dx = data[threei].dx/3.0; + for (j = 1; j < 9; ++j) new_data[index_new + j].dx = new_data[index_new].dx; + + new_data[index_new].x = data[threei].x - new_data[index_new].dx; + new_data[index_new + 1].x = data[threei].x; + new_data[index_new + 2].x = data[threei].x + new_data[index_new].dx; + + new_data[index_new + 3].x = data[threei + 1].x - new_data[index_new].dx; + new_data[index_new + 4].x = data[threei + 1].x; + new_data[index_new + 5].x = data[threei + 1].x + new_data[index_new].dx; + + new_data[index_new + 6].x = data[threei + 2].x - new_data[index_new].dx; + new_data[index_new + 7].x = data[threei + 2].x; + new_data[index_new + 8].x = data[threei + 2].x + new_data[index_new].dx; + + args[arg_to_integ] = new_data[index_new].x; + new_data[index_new].f = function(args, Itable); + new_data[index_new + 1].f = data[threei].f; + args[arg_to_integ] = new_data[index_new + 2].x; + new_data[index_new + 2].f = function(args, Itable); + + args[arg_to_integ] = new_data[index_new + 3].x; + new_data[index_new + 3].f = function(args, Itable); + new_data[index_new + 4].f = data[threei + 1].f; + args[arg_to_integ] = new_data[index_new + 5].x; + new_data[index_new + 5].f = function(args, Itable); + + args[arg_to_integ] = new_data[index_new + 6].x; + new_data[index_new + 6].f = function(args, Itable); + new_data[index_new + 7].f = data[threei + 2].f; + args[arg_to_integ] = new_data[index_new + 8].x; + new_data[index_new + 8].f = function(args, Itable); + + new_abs_d2f_dx[index_new/3] = fabs(new_data[index_new].dx * (new_data[index_new].f - 2.0 * new_data[index_new + 1].f + new_data[index_new + 2].f)); + new_abs_d2f_dx[index_new/3 + 1] = fabs(new_data[index_new].dx * (new_data[index_new + 3].f - 2.0 * new_data[index_new + 4].f + new_data[index_new + 5].f)); + new_abs_d2f_dx[index_new/3 + 2] = fabs(new_data[index_new].dx * (new_data[index_new + 6].f - 2.0 * new_data[index_new + 7].f + new_data[index_new + 8].f)); + + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 1]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 2]); + + integ_res.integ_est += new_data[index_new].dx * (new_data[index_new].f + new_data[index_new + 1].f + new_data[index_new + 2].f + + new_data[index_new + 3].f + new_data[index_new + 4].f + new_data[index_new + 5].f + + new_data[index_new + 6].f + new_data[index_new + 7].f + new_data[index_new + 8].f); + + integ_res.abs_prec += new_abs_d2f_dx[index_new/3] + new_abs_d2f_dx[index_new/3 + 1] + new_abs_d2f_dx[index_new/3 + 2]; + + index_new += 9; + + } // else + + } // for (i < nvals/3) + + integ_res.rel_prec = integ_res.abs_prec/integ_res.integ_est; + + integ_res.n_vals = index_new; + + delete[] data; + data = new_data; + + max_abs_d2f_dx = new_max_abs_d2f_dx; + + delete[] abs_d2f_dx; + abs_d2f_dx = new_abs_d2f_dx; + + return; + } + */ + + Integral_result_CX Integrate_optimal (complex (*function) (Vect_DP), Vect_DP& args, int arg_to_integ, DP xmin, DP xmax, + DP req_rel_prec, DP req_abs_prec, int max_nr_pts) + { + if (xmax < xmin) JSCerror("Use xmax > xmin in Integrate."); + + Integral_data_CX integ_dat (function, args, arg_to_integ, xmin, xmax); + + while ((integ_dat.integ_res.rel_prec > req_rel_prec || integ_dat.integ_res.abs_prec > req_abs_prec) && integ_dat.integ_res.n_vals < max_nr_pts) { + + integ_dat.Improve_estimate (function, args, arg_to_integ, max_nr_pts); + } + + return(integ_dat.integ_res); + } + + // No implementation with I_table yet for complex... + /* + Integral_result Integrate_optimal_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, + I_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts) + { + if (xmax < xmin) JSCerror("Use xmax > xmin in Integrate."); + + Integral_data integ_dat (function, args, arg_to_integ, Itable, xmin, xmax); + + while ((integ_dat.integ_res.rel_prec > req_rel_prec || integ_dat.integ_res.abs_prec > req_abs_prec) && integ_dat.integ_res.n_vals < max_nr_pts) { + + integ_dat.Improve_estimate (function, args, arg_to_integ, Itable, max_nr_pts); + } + + return(integ_dat.integ_res); + } + */ + + // No implementation with I_table yet for complex... + /* + Integral_result Integrate_optimal_using_table (DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, + I_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts, ofstream& outfile) + { + if (xmax < xmin) JSCerror("Use xmax > xmin in Integrate."); + + Integral_data integ_dat (function, args, arg_to_integ, Itable, xmin, xmax); + + while ((integ_dat.integ_res.rel_prec > req_rel_prec || integ_dat.integ_res.abs_prec > req_abs_prec) && integ_dat.integ_res.n_vals < max_nr_pts) { + + integ_dat.Improve_estimate (function, args, arg_to_integ, Itable, max_nr_pts); + + integ_dat.Save (outfile); + + } + + return(integ_dat.integ_res); + } + */ + +} // namespace JSC diff --git a/src/INTEG/Integration_par.cc b/src/INTEG/Integration_par.cc new file mode 100644 index 0000000..d55a959 --- /dev/null +++ b/src/INTEG/Integration_par.cc @@ -0,0 +1,152 @@ +/**************************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) 2006. + +----------------------------------------------------------- + +Integration_par.cc + +Defines all functions to perform integration of functions, parallel (MPI). + +LAST MODIFIED: 30/10/06 + +******************************************************************/ + +#include "JSC.h" +#include "mpi.h" + +using namespace std; + +namespace JSC { + + void Improve_estimate_par (MPI_Comm comm, Integral_data& integdat, DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, I_table Itable, int max_nr_pts) + { + // This function is only ever called by process rank 0. + + // We generate a max of 3* n_vals points + data_pt* new_data = new data_pt[3* integdat.integ_res.n_vals]; + DP* new_abs_d2f_dx = new DP[integdat.integ_res.n_vals]; + + // Check points in batches of 3; if needed, improve + int threei = 0; + int index_new = 0; + int i, j; + + integdat.integ_res.abs_prec = 0.0; + integdat.integ_res.integ_est = 0.0; + + DP new_max_abs_d2f_dx = 0.0; + + for (i = 0; i < integdat.integ_res.n_vals/3; ++i) { + + threei = 3 * i; + + if (integdat.abs_d2f_dx[i] <= 0.1 * integdat.max_abs_d2f_dx || index_new + integdat.integ_res.n_vals - threei > max_nr_pts) { + + // simply transfer the data points into new_data + new_abs_d2f_dx[index_new/3] = integdat.abs_d2f_dx[i]; + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + for (j = 0; j < 3; ++j) { + new_data[index_new].x = integdat.data[threei + j].x; + new_data[index_new].f = integdat.data[threei + j].f; + new_data[index_new].dx = integdat.data[threei + j].dx; + integdat.integ_res.integ_est += new_data[index_new].dx * new_data[index_new].f; + index_new++; + } + integdat.integ_res.abs_prec += abs_d2f_dx[i]; + + } + + else { // create six new entries and transfer the three existing ones + + new_data[index_new].dx = integdat.data[threei].dx/3.0; + for (j = 1; j < 9; ++j) new_data[index_new + j].dx = new_data[index_new].dx; + + new_data[index_new].x = integdat.data[threei].x - new_data[index_new].dx; + new_data[index_new + 1].x = integdat.data[threei].x; + new_data[index_new + 2].x = integdat.data[threei].x + new_data[index_new].dx; + + new_data[index_new + 3].x = integdat.data[threei + 1].x - new_data[index_new].dx; + new_data[index_new + 4].x = integdat.data[threei + 1].x; + new_data[index_new + 5].x = integdat.data[threei + 1].x + new_data[index_new].dx; + + new_data[index_new + 6].x = integdat.data[threei + 2].x - new_data[index_new].dx; + new_data[index_new + 7].x = integdat.data[threei + 2].x; + new_data[index_new + 8].x = integdat.data[threei + 2].x + new_data[index_new].dx; + + args[arg_to_integ] = new_data[index_new].x; + new_data[index_new].f = function(args, Itable); + new_data[index_new + 1].f = integdat.data[threei].f; + args[arg_to_integ] = new_data[index_new + 2].x; + new_data[index_new + 2].f = function(args, Itable); + + args[arg_to_integ] = new_data[index_new + 3].x; + new_data[index_new + 3].f = function(args, Itable); + new_data[index_new + 4].f = integdat.data[threei + 1].f; + args[arg_to_integ] = new_data[index_new + 5].x; + new_data[index_new + 5].f = function(args, Itable); + + args[arg_to_integ] = new_data[index_new + 6].x; + new_data[index_new + 6].f = function(args, Itable); + new_data[index_new + 7].f = integdat.data[threei + 2].f; + args[arg_to_integ] = new_data[index_new + 8].x; + new_data[index_new + 8].f = function(args, Itable); + + new_abs_d2f_dx[index_new/3] = fabs(new_data[index_new].dx * (new_data[index_new].f - 2.0 * new_data[index_new + 1].f + new_data[index_new + 2].f)); + new_abs_d2f_dx[index_new/3 + 1] = fabs(new_data[index_new].dx * (new_data[index_new + 3].f - 2.0 * new_data[index_new + 4].f + new_data[index_new + 5].f)); + new_abs_d2f_dx[index_new/3 + 2] = fabs(new_data[index_new].dx * (new_data[index_new + 6].f - 2.0 * new_data[index_new + 7].f + new_data[index_new + 8].f)); + + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 1]); + new_max_abs_d2f_dx = JSC::max(new_max_abs_d2f_dx, new_abs_d2f_dx[index_new/3 + 2]); + + integdat.integ_res.integ_est += new_data[index_new].dx * (new_data[index_new].f + new_data[index_new + 1].f + new_data[index_new + 2].f + + new_data[index_new + 3].f + new_data[index_new + 4].f + new_data[index_new + 5].f + + new_data[index_new + 6].f + new_data[index_new + 7].f + new_data[index_new + 8].f); + + integdat.integ_res.abs_prec += new_abs_d2f_dx[index_new/3] + new_abs_d2f_dx[index_new/3 + 1] + new_abs_d2f_dx[index_new/3 + 2]; + + index_new += 9; + + } // else + + } // for (i < nvals/3) + + integdat.integ_res.rel_prec = integdat.integ_res.abs_prec/integdat.integ_res.integ_est; + + integdat.integ_res.n_vals = index_new; + + delete[] integdat.data; + integdat.data = new_data; + + integdat.max_abs_d2f_dx = new_max_abs_d2f_dx; + + delete[] integdat.abs_d2f_dx; + integdat.abs_d2f_dx = new_abs_d2f_dx; + + return; + } + + + Integral_result Integrate_optimal_par_using_table (MPI_Comm comm, DP (*function) (Vect_DP, I_table), Vect_DP& args, int arg_to_integ, + I_table Itable, DP xmin, DP xmax, DP req_rel_prec, DP req_abs_prec, int max_nr_pts, ofstream& outfile) + { + if (xmax < xmin) JSCerror("Use xmax > xmin in Integrate."); + + Integral_data integ_dat (function, args, arg_to_integ, Itable, xmin, xmax); + + while ((integ_dat.integ_res.rel_prec > req_rel_prec || integ_dat.integ_res.abs_prec > req_abs_prec) && integ_dat.integ_res.n_vals < max_nr_pts) { + + Improve_estimate_par (comm, integ_dat, function, args, arg_to_integ, Itable, max_nr_pts); + + integ_dat.Save (outfile); + + } + + return(integ_dat.integ_res); + } + + +} // namespace JSC diff --git a/src/LIEBLIN/LiebLin_Bethe_State.cc b/src/LIEBLIN/LiebLin_Bethe_State.cc new file mode 100644 index 0000000..b1f1758 --- /dev/null +++ b/src/LIEBLIN/LiebLin_Bethe_State.cc @@ -0,0 +1,809 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Bethe_State.cc + +Purpose: Definitions for LiebLin_Bethe_State class. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + //*************************************************************************************************** + + // Function definitions: class LiebLin_Bethe_State + + LiebLin_Bethe_State::LiebLin_Bethe_State () + : c_int (0.0), L(0.0), cxL(0.0), N(0), + //OriginStateIxe(Vect(0,1)), + Ix2_available(Vect(0, 1)), index_first_hole_to_right (Vect(0,1)), displacement (Vect(0,1)), + Ix2(Vect(0, 1)), lambdaoc(Vect(0.0, 1)), //BE(Vect(0.0, 1)), + S(Vect(0.0, 1)), dSdlambdaoc(Vect(0.0, 1)), + diffsq(0.0), prec(ITER_REQ_PREC_LIEBLIN), conv(0), iter_Newton(0), E(0.0), iK(0), K(0.0), lnnorm(-100.0) + { + stringstream Nout; Nout << N; label = Nout.str() + LABELSEP + JSCcoding[0] + LABELSEP;//"_0_"; + } + + LiebLin_Bethe_State::LiebLin_Bethe_State (DP c_int_ref, DP L_ref, int N_ref) + : c_int(c_int_ref), L(L_ref), cxL(c_int_ref * L_ref), N(N_ref), + //OriginStateIx2(Vect(0,N), + Ix2_available(Vect(0, 2)), index_first_hole_to_right (Vect(0,N)), displacement (Vect(0,N)), + Ix2(Vect(0, N)), lambdaoc(Vect(0.0, N)), //BE(Vect(0.0, N)), + S(Vect(0.0, N)), dSdlambdaoc(Vect(0.0, N)), + diffsq(0.0), prec(JSC::max(1.0, 1.0/(c_int * c_int)) * ITER_REQ_PREC_LIEBLIN), conv(0), iter_Newton(0), E(0.0), iK(0), K(0.0), lnnorm(-100.0) + { + if (c_int < 0.0) JSCerror("You must use a positive interaction parameter !"); + if (N < 0) JSCerror("Particle number must be strictly positive."); + + stringstream Nout; Nout << N; label = Nout.str() + LABELSEP + JSCcoding[0] + LABELSEP;//+ "_0_"; + + // Set quantum numbers to ground-state configuration: + for (int i = 0; i < N; ++i) Ix2[i] = -(N-1) + 2*i; + + Vect OriginIx2 = Ix2; + + (*this).Set_Label_from_Ix2 (OriginIx2); + //(*this).Set_Label_Internals_from_Ix2 (OriginIx2); + } + + + LiebLin_Bethe_State& LiebLin_Bethe_State::operator= (const LiebLin_Bethe_State& RefState) + { + if (this != &RefState) { + c_int = RefState.c_int; + L = RefState.L; + cxL = RefState.cxL; + N = RefState.N; + label = RefState.label; + Ix2_available = RefState.Ix2_available; + index_first_hole_to_right = RefState.index_first_hole_to_right; + displacement = RefState.displacement; + Ix2 = RefState.Ix2; + lambdaoc = RefState.lambdaoc; + //BE = RefState.BE; + S = RefState.S; + dSdlambdaoc = RefState.dSdlambdaoc; + diffsq = RefState.diffsq; + prec = RefState.prec; + conv = RefState.conv; + iter_Newton = RefState.iter_Newton; + E = RefState.E; + iK = RefState.iK; + K = RefState.K; + lnnorm = RefState.lnnorm; + } + return(*this); + } + + void LiebLin_Bethe_State::Set_to_Label (string label_ref, const Vect& OriginStateIx2) + { + State_Label_Data labeldata = Read_State_Label (label_ref, OriginStateIx2); + + if (N != labeldata.M[0]) { + cout << label_ref << endl; + cout << labeldata.M << endl; + JSCerror("Trying to set an incorrect label on LiebLin_Bethe_State: N != M[0]."); + } + if (N != OriginStateIx2.size()) { + cout << label_ref << endl; + cout << labeldata.M << endl; + JSCerror("Trying to set an incorrect label on LiebLin_Bethe_State: N != OriginStateIx2.size()."); + } + + label = label_ref; + + Vect OriginStateIx2ordered = OriginStateIx2; + OriginStateIx2ordered.QuickSort(); + + // Set all Ix2 to OriginState's Ix2: + for (int i = 0; i < N; ++i) Ix2[i] = OriginStateIx2ordered[i]; + + // Now set the excitations: + for (int iexc = 0; iexc < labeldata.nexc[0]; ++iexc) + for (int i = 0; i < N; ++i) if (Ix2[i] == labeldata.Ix2old[0][iexc]) Ix2[i] = labeldata.Ix2exc[0][iexc]; + + // Now reorder the Ix2 to follow convention: + Ix2.QuickSort(); + + //cout << "Setting label:" << label_ref << endl << "Ix2old = " << labeldata.Ix2old[0] << endl << "Ix2exc = " << labeldata.Ix2exc[0] << endl; + //cout << "on " << OriginStateIx2ordered << endl << "giving " << Ix2 << endl; + + (*this).Set_Label_from_Ix2 (OriginStateIx2ordered); + //(*this).Set_Label_Internals_from_Ix2 (OriginStateIx2ordered); + } + + void LiebLin_Bethe_State::Set_to_Label (string label_ref) + { + // This function assumes that OriginState is the ground state. + + Vect OriginStateIx2(N); + for (int i = 0; i < N; ++i) OriginStateIx2[i] = -N + 1 + 2*i; + + (*this).Set_to_Label(label_ref, OriginStateIx2); + } + + + void LiebLin_Bethe_State::Set_Label_from_Ix2 (const Vect& OriginStateIx2) + { + // This function does not assume any ordering of the Ix2. + + if (N != OriginStateIx2.size()) JSCerror("N != OriginStateIx2.size() in Set_Label_from_Ix2."); + + //cout << "Setting label on Ix2 " << endl << Ix2 << endl; + + // Set the state label: + Vect type_ref(0,1); + Vect M_ref(N, 1); + Vect nexc_ref(0, 1); + // Count nr of particle-holes: + for (int i = 0; i < N; ++i) if (!OriginStateIx2.includes(Ix2[i])) nexc_ref[0] += 1; + Vect > Ix2old_ref(1); + Vect > Ix2exc_ref(1); + Ix2old_ref[0] = Vect(JSC::max(nexc_ref[0],1)); + Ix2exc_ref[0] = Vect(JSC::max(nexc_ref[0],1)); + int nexccheck = 0; + for (int i = 0; i < N; ++i) if (!OriginStateIx2.includes(Ix2[i])) Ix2exc_ref[0][nexccheck++] = Ix2[i]; + if (nexccheck != nexc_ref[0]) JSCerror("Counting excitations wrong (1) in LiebLin_Bethe_State::Set_Label_from_Ix2"); + nexccheck = 0; + for (int i = 0; i < N; ++i) if (!Ix2.includes (OriginStateIx2[i])) Ix2old_ref[0][nexccheck++] = OriginStateIx2[i]; + if (nexccheck != nexc_ref[0]) { + cout << "nexc_ref[0] = " << nexc_ref[0] << "\tnexccheck = " << nexccheck << endl; + cout << OriginStateIx2 << endl; + cout << Ix2 << endl; + cout << nexc_ref[0] << endl; + cout << Ix2old_ref[0] << endl; + cout << Ix2exc_ref[0] << endl; + JSCerror("Counting excitations wrong (2) in LiebLin_Bethe_State::Set_Label_from_Ix2"); + } + // Now order the Ix2old_ref and Ix2exc_ref: + Ix2old_ref[0].QuickSort(); + Ix2exc_ref[0].QuickSort(); + + State_Label_Data labeldata(type_ref, M_ref, nexc_ref, Ix2old_ref, Ix2exc_ref); + + label = Return_State_Label (labeldata, OriginStateIx2); + } + + /* + void LiebLin_Bethe_State::Set_Label_from_Ix2 (const Vect& OriginStateIx2) + { + // This function was deprecated since it assumed that the Ix2 of the state were + // in a particular order mirroring the indices of OriginStateIx2. + + if (N != OriginStateIx2.size()) JSCerror("N != OriginStateIx2.size() in Set_Label_from_Ix2."); + + Vect OriginStateIx2ordered = OriginStateIx2; + OriginStateIx2ordered.QuickSort(); + + // Set the state label: + Vect type_ref(0,1); + Vect M_ref(N, 1); + Vect nexc_ref(0, 1); + // Count nr of particle-holes: + for (int i = 0; i < N; ++i) if (Ix2[i] != OriginStateIx2ordered[i]) nexc_ref[0] += 1; + Vect > Ix2old_ref(1); + Vect > Ix2exc_ref(1); + Ix2old_ref[0] = Vect(JSC::max(nexc_ref[0],1)); + Ix2exc_ref[0] = Vect(JSC::max(nexc_ref[0],1)); + int nexccheck = 0; + for (int i = 0; i < N; ++i) + if (Ix2[i] != OriginStateIx2ordered[i]) { + Ix2old_ref[0][nexccheck] = OriginStateIx2ordered[i]; + Ix2exc_ref[0][nexccheck++] = Ix2[i]; + } + + State_Label_Data labeldata(type_ref, M_ref, nexc_ref, Ix2old_ref, Ix2exc_ref); + + label = Return_State_Label (labeldata); + } + */ + + void LiebLin_Bethe_State::Set_Label_Internals_from_Ix2 (const Vect& OriginStateIx2) + { + //JSCerror("LiebLin_Bethe_State::Set_Label_Internals_from_Ix2 deprecated 20110604"); + + if (N != OriginStateIx2.size()) JSCerror("N != OriginStateIx2.size() in Set_Label_Internals_from_Ix2."); + + Vect OriginStateIx2ordered = OriginStateIx2; + OriginStateIx2ordered.QuickSort(); + + // Set the state label: + Vect type_ref(0,1); + Vect M_ref(N, 1); + Vect nexc_ref(0, 1); + // Count nr of particle-holes: + for (int i = 0; i < N; ++i) if (!OriginStateIx2.includes(Ix2[i])) nexc_ref[0] += 1; + Vect > Ix2old_ref(1); + Vect > Ix2exc_ref(1); + Ix2old_ref[0] = Vect(JSC::max(nexc_ref[0],1)); + Ix2exc_ref[0] = Vect(JSC::max(nexc_ref[0],1)); + int nexccheck = 0; + for (int i = 0; i < N; ++i) + if (Ix2[i] != OriginStateIx2ordered[i]) { + Ix2old_ref[0][nexccheck] = OriginStateIx2ordered[i]; + Ix2exc_ref[0][nexccheck++] = Ix2[i]; + } + + State_Label_Data labeldata(type_ref, M_ref, nexc_ref, Ix2old_ref, Ix2exc_ref); + + label = Return_State_Label (labeldata, OriginStateIx2); + + // Construct the Ix2_available vector: we give one more quantum number on left and right: + int navailable = 2 + (JSC::max(Ix2.max(), OriginStateIx2.max()) - JSC::min(Ix2.min(), OriginStateIx2.min()))/2 - N + 1; + Ix2_available = Vect(navailable); + index_first_hole_to_right = Vect(N); + + // First set Ix2_available to all holes from left + for (int i = 0; i < Ix2_available.size(); ++i) Ix2_available[i] = JSC::min(Ix2.min(), OriginStateIx2.min()) - 2 + 2*i; + + // Now shift according to Ix2 of OriginState: + for (int j = 0; j < N; ++j) { + int i = 0; + while (Ix2_available[i] < OriginStateIx2ordered[j]) i++; + // We now have Ix2_available[i] == OriginStateIx2[j]. Shift all Ix2_available to the right of this by 2; + for (int i1 = i; i1 < navailable; ++i1) Ix2_available[i1] += 2; + index_first_hole_to_right[j] = i; + } + // Ix2_available and index_first_hole_to_right are now fully defined. + + // Now set displacement vector: + displacement = Vect(0, N); + // Set displacement vector from the Ix2: + for (int j = 0; j < N; ++j) { + if (Ix2[j] < OriginStateIx2ordered[j]) { + // Ix2[j] must be equal to some OriginState_Ix2_available[i] for i < OriginState_index_first_hole_to_right[j] + //cout << "Going down\t" << Ix2[j] << "\t" << index_first_hole_to_right[j] << "\t" << displacement[j] << endl; + while (Ix2[j] != Ix2_available[index_first_hole_to_right[j] + displacement[j] ]) { + //cout << j << "\t" << index_first_hole_to_right[j] << "\t" << displacement[j] << "\t" << Ix2_available[index_first_hole_to_right[j] + displacement[j] ] << endl; + if (index_first_hole_to_right[j] + displacement[j] == 0) { + cout << label << endl << j << endl << OriginStateIx2 << endl << Ix2 << endl << Ix2_available << endl << index_first_hole_to_right << endl << displacement << endl; + JSCerror("Going down too far in Set_Label_Internals..."); + } + displacement[j]--; + } + } + if (Ix2[j] > OriginStateIx2ordered[j]) { + // Ix2[j] must be equal to some Ix2_available[i] for i >= index_first_hole_to_right[j] + //cout << "Going up\t" << Ix2[j] << "\t" << index_first_hole_to_right[j] << "\t" << displacement[j] << endl; + displacement[j] = 1; // start with this value to prevent segfault + while (Ix2[j] != Ix2_available[index_first_hole_to_right[j] - 1 + displacement[j] ]) { + //cout << j << "\t" << index_first_hole_to_right[j] << "\t" << displacement[j] << "\t" << Ix2_available[index_first_hole_to_right[j] + displacement[j] ] << endl; + if (index_first_hole_to_right[j] + displacement[j] == Ix2_available.size() - 1) { + cout << label << endl << j << endl << OriginStateIx2 << endl << Ix2 << endl << Ix2_available << endl << index_first_hole_to_right << endl << displacement << endl; + JSCerror("Going up too far in Set_Label_Internals..."); + } + displacement[j]++; + } + } + } + //cout << "label " << label << endl; + //cout << "Ix2: " << Ix2 << endl << "Ix2_available: " << Ix2_available << endl << "index...: " << index_first_hole_to_right << endl << "displacement: " << displacement << endl; + //char a; cin >> a; + + + } + + bool LiebLin_Bethe_State::Check_Admissibility (char whichDSF) + { + //if (Ix2.min() < -13 || Ix2.max() > 13) return(false); // For testing with restricted Hilbert space + return(true); + } + + void LiebLin_Bethe_State::Find_Rapidities (bool reset_rapidities) + { + // This function finds the rapidities of the eigenstate + + lnnorm = -100.0; // sentinel value, recalculated if Newton method used in the last step of iteration. + + diffsq = 1.0; + + if (reset_rapidities) (*this).Set_Free_lambdaocs(); + /* + // Start with normal iterations: + //for (int niternormal = 0; niternormal < 10; ++niternormal) { + for (int niternormal = 0; niternormal < N; ++niternormal) { + //(*this).Iterate_BAE(0.99); + (*this).Iterate_BAE(0.9); + cout << "Normal: " << niternormal << "\t" << diffsq << endl; + cout << (*this).lambdaoc << endl; + if (diffsq < sqrt(prec)) break; + } + */ + iter_Newton = 0; + + DP damping = 1.0; + DP diffsq_prev = 1.0e+6; + //while (diffsq > prec && !is_nan(diffsq) && iter_Newton < 40) { + while (diffsq > prec && !is_nan(diffsq) && iter_Newton < 100) { + //while (diffsq > prec && !is_nan(diffsq) && iter_Newton < 400) { + (*this).Iterate_BAE_Newton(damping); + //(*this).Iterate_BAE_S(damping); // Not as fast as Newton for N up to ~ 256 + //if (diffsq > diffsq_prev) damping /= 2.0; + if (diffsq > diffsq_prev && damping > 0.5) damping /= 2.0; + else if (diffsq < diffsq_prev) damping = 1.0; + diffsq_prev = diffsq; + //cout << iter_Newton << "\t" << diffsq << "\t" << damping << endl; + //cout << (*this).lambdaoc << endl; + } + + conv = ((diffsq < prec) && (*this).Check_Rapidities()) ? 1 : 0; + + if (!conv) { + cout << "Alert! State " << label << " did not converge... diffsq " << diffsq << "\titer_Newton " << iter_Newton << (*this) << endl; + //char a; cin >> a; + } + + //cout << (*this) << endl; + + return; + } + + bool LiebLin_Bethe_State::Check_Rapidities() + { + bool nonan = true; + + for (int j = 0; j < N; ++j) nonan *= !is_nan(lambdaoc[j]); + + return nonan; + } + + DP LiebLin_Bethe_State::String_delta() + { + return(0.0); // no strings (thus no deviations) in replusive LiebLin + } + + bool LiebLin_Bethe_State::Check_Symmetry () + { + // Checks whether the I's are symmetrically distributed. + + bool symmetric_state = true; + + Vect Ix2check = Ix2; + Ix2check.QuickSort(); + + for (int alpha = 0; alpha <= N/2; ++alpha) + symmetric_state = symmetric_state && (Ix2check[alpha] == -Ix2check[N - 1 - alpha]); + + return(symmetric_state); + } + + void LiebLin_Bethe_State::Compute_lnnorm () + { + if (lnnorm == -100.0) { // else Gaudin part already calculated by Newton method + + SQMat_DP Gaudin_Red(N); + + (*this).Build_Reduced_Gaudin_Matrix(Gaudin_Red); + + lnnorm = real(lndet_LU_dstry(Gaudin_Red)); + + // Add the pieces outside of Gaudin determinant + + for (int j = 0; j < N - 1; ++j) for (int k = j+1; k < N; ++k) lnnorm += log(1.0 + 1.0/pow(lambdaoc[j] - lambdaoc[k], 2.0)); + + } + + return; + } + + void LiebLin_Bethe_State::Compute_All (bool reset_rapidities) // solves BAE, computes E, K and lnnorm + { + (*this).Find_Rapidities (reset_rapidities); + if (conv == 1) { + (*this).Compute_Energy (); + (*this).Compute_Momentum (); + (*this).Compute_lnnorm (); + } + return; + } + + void LiebLin_Bethe_State::Set_Free_lambdaocs() + { + if (cxL >= 1.0) + for (int a = 0; a < N; ++a) lambdaoc[a] = PI * Ix2[a]/cxL; + + // For small values of c, use better approximation using approximate zeroes of Hermite polynomials: see Gaudin eqn 4.71. + if (cxL < 1.0) { + //DP sqrtcL = pow(cxL, 0.5); + DP oneoversqrtcLN = 1.0/pow(cxL * N, 0.5); + for (int a = 0; a < N; ++a) lambdaoc[a] = oneoversqrtcLN * PI * Ix2[a]; + + //for (int a = 0; a < N; ++a) lambdaoc[a] = sqrtcL * PI * Ix2[a]; // wrong values, correct scaling with c + //for (int a = 0; a < N; ++a) lambdaoc[a] = PI * Ix2[a]/(cxL + 2.0 * N); // set to minimal distance lattice + } + + //cout << "Set free lambdaocs to: " << endl << lambdaoc << endl; + return; + } + + void LiebLin_Bethe_State::Iterate_BAE (DP damping) + { + // does one step of simple iterations + + DP sumtheta = 0.0; + Vect_DP dlambdaoc (0.0, N); + + for (int j = 0; j < N; ++j) { + + sumtheta = 0.0; + for (int k = 0; k < N; ++k) sumtheta += atan((lambdaoc[j] - lambdaoc[k])); + sumtheta *= 2.0; + + dlambdaoc[j] = damping * ((PI*Ix2[j] - sumtheta)/cxL - lambdaoc[j]); + + } + + diffsq = 0.0; + //DP sumsq = 0.0; + for (int i = 0; i < N; ++i) { + lambdaoc[i] += dlambdaoc[i]; + //sumsq += lambdaoc[i] * lambdaoc[i]; + //diffsq += dlambdaoc[i] * dlambdaoc[i]; + //if (cxL > 1.0) diffsq += dlambdaoc[i] * dlambdaoc[i]; + //else diffsq += cxL * cxL * dlambdaoc[i] * dlambdaoc[i]; + // Normalize the diffsq by the typical value of the rapidities: + if (cxL > 1.0) diffsq += dlambdaoc[i] * dlambdaoc[i]/(lambdaoc[i] * lambdaoc[i] + 1.0e-6); + else diffsq += cxL * cxL * dlambdaoc[i] * dlambdaoc[i]/(lambdaoc[i] * lambdaoc[i] + 1.0e-6); + } + diffsq /= DP(N); + //diffsq /= sqrt(sumsq) * DP(N); + + return; + } + + void LiebLin_Bethe_State::Iterate_BAE_S (DP damping) + { + // This is essentially Newton's method but only in one variable. + // The logic is that the derivative of the LHS of the BE_j w/r to lambdaoc_j is much larger than with respect to lambdaoc_l with l != j. + + Vect_DP dlambdaoc (0.0, N); + + // Start by calculating S and dSdlambdaoc: + for (int j = 0; j < N; ++j) { + + S[j] = 0.0; + for (int k = 0; k < N; ++k) S[j] += atan((lambdaoc[j] - lambdaoc[k])); + S[j] *= 2.0/cxL; + + dSdlambdaoc[j] = 0.0; + for (int k = 0; k < N; ++k) dSdlambdaoc[j] += 1.0/((lambdaoc[j] - lambdaoc[k]) * (lambdaoc[j] - lambdaoc[k]) + 1.0); + dSdlambdaoc[j] *= 2.0/(PI * cxL); + + dlambdaoc[j] = (PI*Ix2[j]/cxL - S[j] + lambdaoc[j] * dSdlambdaoc[j])/(1.0 + dSdlambdaoc[j]) - lambdaoc[j]; + + } + + diffsq = 0.0; + for (int i = 0; i < N; ++i) { + lambdaoc[i] += damping * dlambdaoc[i]; + // Normalize the diffsq by the typical value of the rapidities: + if (cxL > 1.0) diffsq += dlambdaoc[i] * dlambdaoc[i]/(lambdaoc[i] * lambdaoc[i] + 1.0e-6); + else diffsq += cxL * cxL * dlambdaoc[i] * dlambdaoc[i]/(lambdaoc[i] * lambdaoc[i] + 1.0e-6); + } + diffsq /= DP(N); + + iter_Newton++; + + return; + } + + void LiebLin_Bethe_State::Iterate_BAE_Newton (DP damping) + { + // does one step of a Newton method on the rapidities... + + Vect_DP RHSBAE (0.0, N); // contains RHS of BAEs + Vect_DP dlambdaoc (0.0, N); // contains delta lambdaoc computed from Newton's method + SQMat_DP Gaudin (0.0, N); + Vect_INT indx (N); + DP sumtheta = 0.0; + int atanintshift = 0; // for large |lambda|, use atan (lambda) = sgn(lambda) pi/2 - atan(1/lambda) + DP lambdahere = 0.0; + + // Compute the RHS of the BAEs: + + for (int j = 0; j < N; ++j) { + + sumtheta = 0.0; + atanintshift = 0; + for (int k = 0; k < N; ++k) + if (j != k) { // otherwise 0 + if (fabs(lambdahere = lambdaoc[j] - lambdaoc[k]) < 1.0) { // use straight atan + sumtheta += atan(lambdahere); + } + else { // for large rapidities, use dual form of atan, extracting pi/2 factors + atanintshift += sgn_DP(lambdahere); + sumtheta -= atan(1.0/lambdahere); + } + } + sumtheta *= 2.0; + + //cout << j << "\t" << Ix2[j] << "\t" << atanintshift << endl; + + RHSBAE[j] = cxL * lambdaoc[j] + sumtheta - PI*(Ix2[j] - atanintshift); + } + + (*this).Build_Reduced_Gaudin_Matrix (Gaudin); + + for (int j = 0; j < N; ++j) dlambdaoc[j] = - RHSBAE[j]; + + //cout << "dlambdaoc pre lubksb = " << dlambdaoc << endl; + + DP d; + ludcmp (Gaudin, indx, d); + lubksb (Gaudin, indx, dlambdaoc); + + //cout << "dlambdaoc post lubksb = " << dlambdaoc << endl; + + bool ordering_changed = false; + for (int j = 0; j < N-1; ++j) if (lambdaoc[j] + dlambdaoc[j] > lambdaoc[j+1] + dlambdaoc[j+1]) ordering_changed = true; + + // To prevent Newton from diverging, we limit the size of the rapidity changes. + // The leftmost and rightmost rapidities can grow by one order of magnitude per iteration step. + if (ordering_changed) { // We explicitly ensure that the ordering remains correct after the iteration step. + bool ordering_still_changed = false; + DP maxdlambdaoc = 0.0; + do { + ordering_still_changed = false; + if (dlambdaoc[0] < 0.0 && fabs(dlambdaoc[0]) > (maxdlambdaoc = 10.0*JSC::max(fabs(lambdaoc[0]), fabs(lambdaoc[N-1])))) + dlambdaoc[0] = -maxdlambdaoc; + if (lambdaoc[0] + dlambdaoc[0] > lambdaoc[1] + dlambdaoc[1]) { + dlambdaoc[0] = 0.25 * (lambdaoc[1] + dlambdaoc[1] - lambdaoc[0] ); // max quarter distance + ordering_still_changed = true; + //cout << "reason 1" << endl; + } + if (dlambdaoc[N-1] > 0.0 && fabs(dlambdaoc[N-1]) > (maxdlambdaoc = 10.0*JSC::max(fabs(lambdaoc[0]), fabs(lambdaoc[N-1])))) + dlambdaoc[N-1] = maxdlambdaoc; + if (lambdaoc[N-1] + dlambdaoc[N-1] < lambdaoc[N-2] + dlambdaoc[N-2]) { + dlambdaoc[N-1] = 0.25 * (lambdaoc[N-2] + dlambdaoc[N-2] - lambdaoc[N-1]); + ordering_still_changed = true; + //cout << "reason 2" << endl; + } + for (int j = 1; j < N-1; ++j) { + if (lambdaoc[j] + dlambdaoc[j] > lambdaoc[j+1] + dlambdaoc[j+1]) { + dlambdaoc[j] = 0.25 * (lambdaoc[j+1] + dlambdaoc[j+1] - lambdaoc[j]); + ordering_still_changed = true; + //cout << "reason 3" << endl; + } + if (lambdaoc[j] + dlambdaoc[j] < lambdaoc[j-1] + dlambdaoc[j-1]) { + dlambdaoc[j] = 0.25 * (lambdaoc[j-1] + dlambdaoc[j-1] - lambdaoc[j]); + ordering_still_changed = true; + //cout << "reason 4" << endl; + } + } + } while (ordering_still_changed); + } + + //if (ordering_changed) cout << "dlambdaoc post checking = " << dlambdaoc << endl; + + /* + bool orderingchanged = false; + //bool dlambdaexceedsmaxrapdiff = false; + DP damping_to_use = damping; + for (int j = 0; j < N-1; ++j) { + if (lambdaoc[j] + dlambdaoc[j] > lambdaoc[j+1] + dlambdaoc[j+1]) // unacceptable, order changed! + orderingchanged = true; + // We must damp the dlambda such that the ordering is unchanged: + // the condition is lambdaoc[j] + damping*dlambdaoc[j] < lambdaoc[j+1] + damping*dlambdaoc[j+1] + damping_to_use = JSC::min(damping_to_use, 0.5 * (lambdaoc[j+1] - lambdaoc[j])/(dlambdaoc[j] - dlambdaoc[j+1])); + } + */ + //for (int j = 0; j < N; ++j) + //if (fabs(dlambdaoc[j]) > maxrapdiff) dlambdaexceedsmaxrapdiff = true; + /* + // The dlambdaoc must be smaller than the distance to neighbouring rapidities: + // If any dlambdaoc is greater than half the distance, set all to half the previous distance: + if (orderingchanged || dlambdaexceedsmaxrapdiff) { + if (dlambdaoc[0] > 0.0) dlambdaoc[0] = 0.25 * (lambdaoc[1] - lambdaoc[0]); + //else dlambdaoc[0] = 0.25 * (lambdaoc[0] - lambdaoc[1]); + else dlambdaoc[0] = -fabs(lambdaoc[0]); // strictly limit the growth of lambdaoc[0] + //if (dlambdaoc[N-1] > 0.0) dlambdaoc[N-1] = 0.25 * (lambdaoc[N-1] - lambdaoc[N-2]); + if (dlambdaoc[N-1] > 0.0) dlambdaoc[N-1] = fabs(lambdaoc[N-1]); // strictly limit the growth of lambdaoc[N-1] + else dlambdaoc[N-1] = 0.25 * (lambdaoc[N-2] - lambdaoc[N-1]); + for (int j = 1; j < N-1; ++j) { + if (dlambdaoc[j] > 0.0) dlambdaoc[j] = 0.25 * (lambdaoc[j+1] - lambdaoc[j]); + if (dlambdaoc[j] < 0.0) dlambdaoc[j] = 0.25 * (lambdaoc[j-1] - lambdaoc[j]); + } + //cout << "Corrected dlambdaoc = " << dlambdaoc << endl; + } + */ + diffsq = 0.0; + for (int i = 0; i < N; ++i) { + //diffsq += dlambdaoc[i] * dlambdaoc[i]; + // Normalize the diffsq by the typical value of the rapidities: + if (cxL > 1.0) diffsq += dlambdaoc[i] * dlambdaoc[i]/(lambdaoc[i] * lambdaoc[i] + 1.0e-6); + else diffsq += cxL * cxL * dlambdaoc[i] * dlambdaoc[i]/(lambdaoc[i] * lambdaoc[i] + 1.0e-6); + } + diffsq /= DP(N); + + if (ordering_changed) diffsq = 1.0; // reset if Newton wanted to change ordering + + for (int j = 0; j < N; ++j) lambdaoc[j] += damping * dlambdaoc[j]; + //for (int j = 0; j < N; ++j) lambdaoc[j] += damping_to_use * dlambdaoc[j]; + + iter_Newton++; + + // if we've converged, calculate the norm here, since the work has been done... + + //if (diffsq < prec && !orderingchanged && !dlambdaexceedsmaxrapdiff) { + if (diffsq < prec && !ordering_changed) { + + lnnorm = 0.0; + for (int j = 0; j < N; j++) lnnorm += log(fabs(Gaudin[j][j])); + + // Add the pieces outside of Gaudin determinant + for (int j = 0; j < N - 1; ++j) for (int k = j+1; k < N; ++k) lnnorm += log(1.0 + 1.0/pow(lambdaoc[j] - lambdaoc[k], 2.0)); + } + + return; + } + + void LiebLin_Bethe_State::Compute_Energy() + { + E = 0.0; + for (int j = 0; j < N; ++j) E += lambdaoc[j] * lambdaoc[j]; + E *= c_int * c_int; + } + + void LiebLin_Bethe_State::Compute_Momentum() + { + iK = 0; + for (int j = 0; j < N; ++j) { + //cout << j << "\t" << iK << "\t" << Ix2[j] << endl; + iK += Ix2[j]; + } + if (iK % 2) { + cout << Ix2 << endl; + cout << iK << "\t" << iK % 2 << endl; + JSCerror("Sum of Ix2 is not even: inconsistency."); + } + iK /= 2; // sum of Ix2 is guaranteed even. + + K = 2.0 * iK * PI/L; + } + + DP LiebLin_Bethe_State::Kernel (int a, int b) + { + return(2.0/(pow(lambdaoc[a] - lambdaoc[b], 2.0) + 1.0)); + } + + DP LiebLin_Bethe_State::Kernel (DP lambdaoc_ref) + { + return(2.0/(lambdaoc_ref * lambdaoc_ref + 1.0)); + } + + void LiebLin_Bethe_State::Build_Reduced_Gaudin_Matrix (SQMat& Gaudin_Red) + { + + if (Gaudin_Red.size() != N) JSCerror("Passing matrix of wrong size in Build_Reduced_Gaudin_Matrix."); + + DP sum_Kernel = 0.0; + + for (int j = 0; j < N; ++j) + for (int k = 0; k < N; ++k) { + + if (j == k) { + sum_Kernel = 0.0; + for (int kp = 0; kp < N; ++kp) if (j != kp) sum_Kernel += Kernel (lambdaoc[j] - lambdaoc[kp]); + Gaudin_Red[j][k] = cxL + sum_Kernel; + } + + else Gaudin_Red[j][k] = - Kernel (lambdaoc[j] - lambdaoc[k]); + + } + + return; + } + + void LiebLin_Bethe_State::Build_Reduced_BEC_Quench_Gaudin_Matrix (SQMat& Gaudin_Red) + { + // Passing a matrix of dimension N/2 + + if (N % 2 != 0) JSCerror("Choose a state with even numer of particles please"); + + // Check Parity invariant + + bool ck = true; + + for (int j = 0; j < N/2; ++j){ if(Ix2[j] != - Ix2[N-j-1]) ck = false;} + + if (!ck) JSCerror("Choose a parity invariant state please"); + + if (Gaudin_Red.size() != N/2) JSCerror("Passing matrix of wrong size in Build_Reduced_Gaudin_Matrix."); + + DP sum_Kernel = 0.0; + + for (int j = 0; j < N/2; ++j) + for (int k = 0; k < N/2; ++k) { + + if (j == k) { + sum_Kernel = 0.0; + for (int kp = N/2; kp < N; ++kp) if (j + N/2 != kp) sum_Kernel += Kernel (lambdaoc[j+N/2] - lambdaoc[kp]) + Kernel (lambdaoc[j+N/2] + lambdaoc[kp]); + Gaudin_Red[j][k] = cxL + sum_Kernel; + } + + else Gaudin_Red[j][k] = - (Kernel (lambdaoc[j+ N/2] - lambdaoc[k+ N/2]) + Kernel (lambdaoc[j+ N/2] + lambdaoc[k+ N/2]) ); + + } + + return; + } + + + void LiebLin_Bethe_State::Annihilate_ph_pair (int ipart, int ihole, const Vect& OriginStateIx2) + { + // This function changes the Ix2 of a given state by annihilating a particle and hole + // pair specified by ipart and ihole (counting from the left, starting with index 0). + + //cout << "Annihilating ipart " << ipart << " and ihole " << ihole << " of state with label " << (*this).label << endl; + + State_Label_Data currentdata = Read_State_Label ((*this).label, OriginStateIx2); + //cout << "current Ix2old " << currentdata.Ix2old[0] << endl; + //cout << "current Ix2exc " << currentdata.Ix2exc[0] << endl; + + if (ipart >= currentdata.nexc[0]) JSCerror("Particle label too large in LiebLin_Bethe_State::Annihilate_ph_pair."); + if (ihole >= currentdata.nexc[0]) JSCerror("Hole label too large in LiebLin_Bethe_State::Annihilate_ph_pair."); + + // Simply remove the given pair: + Vect type_new = currentdata.type; + Vect M_new = currentdata.M; + Vect nexc_new = currentdata.nexc; + nexc_new[0] -= 1; // we drill one more particle-hole pair at level 0 + int ntypespresent = 1; // only one type for LiebLin + Vect > Ix2old_new(ntypespresent); + Vect > Ix2exc_new(ntypespresent); + for (int it = 0; it < ntypespresent; ++it) Ix2old_new[it] = Vect(JSC::max(nexc_new[it],1)); + for (int it = 0; it < ntypespresent; ++it) Ix2exc_new[it] = Vect(JSC::max(nexc_new[it],1)); + + // Copy earlier data in, leaving out ipart and ihole: + for (int it = 0; it < ntypespresent; ++it) { + for (int i = 0; i < nexc_new[it]; ++i) { + Ix2old_new[it][i] = currentdata.Ix2old[it][i + (i >= ihole)]; + Ix2exc_new[it][i] = currentdata.Ix2exc[it][i + (i >= ipart)]; + } + } + //cout << "Ix2old_new " << Ix2old_new[0] << endl; + //cout << "Ix2exc_new " << Ix2exc_new[0] << endl; + + State_Label_Data newdata (type_new, M_new, nexc_new, Ix2old_new, Ix2exc_new); + + (*this).Set_to_Label (Return_State_Label(newdata, OriginStateIx2)); + + //cout << "Obtained label " << (*this).label << endl; + } + + void LiebLin_Bethe_State::Parity_Flip () + { + // For simplicity, we don't redo base_id, type_id, id. + Vect_INT Ix2buff = Ix2; + Vect_DP lambdaocbuff = lambdaoc; + for (int i = 0; i < N; ++i) Ix2[i] = -Ix2buff[N - 1 - i]; + for (int i = 0; i < N; ++i) lambdaoc[i] = -lambdaocbuff[N - 1 - i]; + iK = -iK; + K = -K; + } + + std::ostream& operator<< (std::ostream& s, const LiebLin_Bethe_State& state) + { + s << endl << "******** State for c = " << state.c_int << " L = " << state.L << " N = " << state.N + << " with label " << state.label << " ********" << endl; + s << "Ix2:" << endl; + for (int j = 0; j < state.N; ++j) s << state.Ix2[j] << " "; + s << endl << "lambdaocs:" << endl; + for (int j = 0; j < state.N; ++j) s << state.lambdaoc[j] << " "; + s << endl << "conv = " << state.conv << " iter_Newton = " << state.iter_Newton << endl; + s << "E = " << state.E << " iK = " << state.iK << " K = " << state.K << " lnnorm = " << state.lnnorm << endl; + + return(s); + } + + +} // namespace JSC diff --git a/src/LIEBLIN/LiebLin_Chem_Pot.cc b/src/LIEBLIN/LiebLin_Chem_Pot.cc new file mode 100644 index 0000000..82c7883 --- /dev/null +++ b/src/LIEBLIN/LiebLin_Chem_Pot.cc @@ -0,0 +1,38 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/LIEBLIN/LiebLin_Chem_Pot.cc + +Purpose: calculates the chemical potential. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + DP Chemical_Potential (LiebLin_Bethe_State& RefState) + { + // RefState is used here to provide the c_int, L and N parameters. + + LiebLin_Bethe_State Nplus1State(RefState.c_int, RefState.L, RefState.N + 1); + //LiebLin_Bethe_State NState(RefState.c_int, RefState.L, RefState.N); + LiebLin_Bethe_State Nmin1State(RefState.c_int, RefState.L, RefState.N - 1); + + Nplus1State.Compute_All(true); + //NState.Compute_All(true); + Nmin1State.Compute_All(true); + + //return(NState.E - Nmin1State.E); + return(0.5 * (Nplus1State.E - Nmin1State.E)); + } + +} // namespace JSC diff --git a/src/LIEBLIN/LiebLin_Matrix_Element_Contrib.cc b/src/LIEBLIN/LiebLin_Matrix_Element_Contrib.cc new file mode 100644 index 0000000..7f7ee52 --- /dev/null +++ b/src/LIEBLIN/LiebLin_Matrix_Element_Contrib.cc @@ -0,0 +1,184 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/LIEBLIN/LiebLin_Matrix_Element_Contrib.cc + +Purpose: handles the generic call for a Lieb-Liniger matrix element contribution. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + //DP Compute_Matrix_Element_Contrib (char whichDSF, bool fixed_iK, LiebLin_Bethe_State& LeftState, + // LiebLin_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile) + //DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, LiebLin_Bethe_State& LeftState, + // LiebLin_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile) + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, LiebLin_Bethe_State& LeftState, + LiebLin_Bethe_State& RefState, DP Chem_Pot, stringstream& DAT_outfile) + { + // This function prints the matrix element information to the fstream, + // and returns a weighed `data_value' to be multiplied by sumrule_factor, + // to determine if scanning along this thread should be continued. + + bool fixed_iK = (iKmin == iKmax); + + // Identify which matrix element is needed from the number of particles in Left and Right states: + DP ME = 0.0; + complex ME_CX = 0.0; + + if (whichDSF == 'Z') + ME = LeftState.E - RefState.E; + else if (whichDSF == 'd' || whichDSF == '1') + ME = real(exp(ln_Density_ME (LeftState, RefState))); + else if (whichDSF == 'o') + ME = real(exp(ln_Psi_ME (LeftState, RefState))); + else if (whichDSF == 'g') + ME = real(exp(ln_Psi_ME (RefState, LeftState))); + else if (whichDSF == 'q') // geometrical quench + //ME_CX = (LiebLin_Twisted_ln_Overlap(1.0, RefState.lambdaoc, RefState.lnnorm, LeftState)); + //ME_CX = (LiebLin_ln_Overlap(RefState.lambdaoc, RefState.lnnorm, LeftState)); + ME_CX = LiebLin_ln_Overlap(LeftState.lambdaoc, LeftState.lnnorm, RefState); + else if (whichDSF == 'B') { // BEC to finite c quench: g2(x=0) + //ME_CX = exp(ln_Overlap_with_BEC (LeftState)); // overlap part + ME_CX = exp(ln_Overlap_with_BEC (LeftState) - ln_Overlap_with_BEC (RefState)); // overlap part, relative to saddle-point state + ME = real(exp(ln_g2_ME (RefState, LeftState))); // matrix element part + // The product is ME_CX * ME = e^{-\delta S_e} \langle \rho_{sp} | g2 (x=0) | \rho_{sp} + e \rangle + // and is the first half of the t-dep expectation value formula coming from the Quench Action formalism + } + else if (whichDSF == 'C') { // BEC to finite c quench: overlap + ME_CX = exp(2.0* ln_Overlap_with_BEC (LeftState)); // overlap sq part + ME = 0.0; + } + else JSCerror("Wrong whichDSF in Compute_Matrix_Element_Contrib."); + + if (is_nan(ME)) ME = (whichDSF == 'Z') ? 1.0e+200 : 0.0; + if (is_nan(norm(ME_CX))) ME_CX = -100.0; + + + // Print information to fstream: + if (LeftState.iK - RefState.iK >= iKmin && LeftState.iK - RefState.iK <= iKmax) { + if (whichDSF == 'Z') { + DAT_outfile << endl << LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot << "\t" + << LeftState.iK - RefState.iK + //<< "\t" << LeftState.conv + << 0 << "\t" // This is the deviation, here always 0 + << "\t" << LeftState.label; + } + else if (whichDSF == 'q') { + DAT_outfile << endl << LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot << "\t" + << LeftState.iK - RefState.iK << "\t" + << real(ME_CX) << "\t" << imag(ME_CX) - twoPI * int(imag(ME_CX)/twoPI + 1.0e-10) << "\t" + //<< LeftState.conv << "\t" + << 0 << "\t" // This is the deviation, here always 0 + << LeftState.label; + } + else if (whichDSF == '1') { + DAT_outfile << endl << LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot << "\t" + << LeftState.iK - RefState.iK << "\t" + << ME << "\t" + //<< LeftState.conv << "\t" + << 0 << "\t" // This is the deviation, here always 0 + << LeftState.label; + DAT_outfile << "\t" << Momentum_Right_Excitations(LeftState) << "\t" << Momentum_Left_Excitations(LeftState); + //cout << Momentum_Right_Excitations(LeftState) << "\t" << Momentum_Left_Excitations(LeftState); + } + else if (whichDSF == 'B') { // BEC to finite c > 0 quench; g2 (x=0) + if (fabs(real(ME_CX) * ME) > 1.0e-100) + DAT_outfile << endl << LeftState.E - RefState.E << "\t" + << LeftState.iK - RefState.iK << "\t" + << real(ME_CX) << "\t" // the overlap is always real + << ME << "\t" + //<< 0 << "\t" // This is the deviation, here always 0 // omit this here + << LeftState.label; + } + else if (whichDSF == 'C') { // BEC to finite c, overlap sq + if (fabs(real(ME_CX)) > 1.0e-100) + DAT_outfile << endl << LeftState.E - RefState.E << "\t" + << LeftState.iK - RefState.iK << "\t" + << real(ME_CX) << "\t" // the overlap is always real + << ME << "\t" + //<< 0 << "\t" // This is the deviation, here always 0 // omit this here + << LeftState.label; + } + else { + DAT_outfile << endl << LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot << "\t" + << LeftState.iK - RefState.iK << "\t" + << ME << "\t" + //<< LeftState.conv << "\t" + << 0 << "\t" // This is the deviation, here always 0 + << LeftState.label; + } + } // if iKmin <= iK <= iKmax + + // Calculate and return the data_value: + DP data_value = ME * ME; + if (whichDSF == 'Z') // use 1/(1 + omega) + data_value = 1.0/(1.0 + LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot); + else if (whichDSF == 'd' || whichDSF == '1') { + if (fixed_iK) + /* + // use omega * MEsq/iK^2 + //data_value = (LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot) + // MEsq/JSC::max(1, (LeftState.iK - RefState.iK) * (LeftState.iK - RefState.iK)) + //: 0.0; + */ + // Careful: use fabs(E) since this must also work with Tgt0 or arbitrary RefState DEPRECATED ++G_1, USE abs_data_value + data_value = (LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot) * ME * ME; + else if (!fixed_iK) // use ME if momentum in fundamental window -iK_UL <= iK <= iK_UL + //data_value = abs(LeftState.iK - RefState.iK) <= RefState.Tableau[0].Ncols ? // this last nr is iK_UL + //MEsq : 0.0; + //data_value = (LeftState.iK - RefState.iK == 0 ? 1.0 : 2.0) * MEsq; + //data_value = ME * ME; + // use omega * MEsq/iK^2 + // Careful: use fabs(E) since this must also work with Tgt0 or arbitrary RefState DEPRECATED ++G_1, USE abs_data_value + data_value = (LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot) * ME * ME/JSC::max(1, (LeftState.iK - RefState.iK) * (LeftState.iK - RefState.iK)); + } + else if (whichDSF == 'g' || whichDSF == 'o') { + if (fixed_iK) + /* + // use omega * MEsq/((k^2 - mu + 4 c N/L)/L) + data_value = (LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot) + * MEsq/((((LeftState.K - RefState.K) * (LeftState.K - RefState.K) - Chem_Pot) + 4.0 * RefState.c_int * RefState.N/RefState.L)/RefState.L) + : 0.0; + */ + // Careful: use fabs(E) since this must also work with Tgt0 or arbitrary RefState + data_value = (LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot) * ME * ME; + else if (!fixed_iK) { // simply use MEsq if momentum in fundamental window -iK_UL <= iK <= iK_UL + //data_value = abs(LeftState.iK - RefState.iK) <= RefState.Tableau[0].Ncols ? // this last nr is iK_UL + //MEsq : 0.0; + //data_value = (LeftState.iK - RefState.iK == 0 ? 1.0 : 2.0) * MEsq; + data_value = ME * ME; + //data_value = fabs(LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot) * ME * ME; + // In the case of whichDSF == 'o', the state with label (N-1)_0_ gives zero data_value. Force to 1. + //if (whichDSF == 'o' && fabs(LeftState.E - RefState.E - (LeftState.N - RefState.N) * Chem_Pot) < 1.0e-10) data_value = 1.0; + } + } + //else if (whichDSF == 'o') + //data_value = abs(LeftState.iK - RefState.iK) <= RefState.Tableau[0].Ncols ? // this last nr is iK_UL + //MEsq : 0.0; + else if (whichDSF == 'q') + data_value = exp(2.0 * real(ME_CX)); + else if (whichDSF == 'B') + data_value = abs(ME_CX * ME)/(1.0 + sqrt(fabs(LeftState.E - RefState.E))); + else if (whichDSF == 'C') + data_value = abs(ME_CX); + + if (whichDSF == '1') // hard cutoff for nr, nl when calculating exponents + if (Momentum_Right_Excitations(LeftState) >= 30 || Momentum_Left_Excitations(LeftState) >= 30) + data_value = 0.0; + + return(data_value); + } + +} // namespace JSC diff --git a/src/LIEBLIN/LiebLin_State_Ensemble.cc b/src/LIEBLIN/LiebLin_State_Ensemble.cc new file mode 100644 index 0000000..9ad20d4 --- /dev/null +++ b/src/LIEBLIN/LiebLin_State_Ensemble.cc @@ -0,0 +1,929 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_State_Ensemble.cc + +Purpose: State ensembles for Lieb-Liniger + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + // Constructors: + + LiebLin_Diagonal_State_Ensemble::LiebLin_Diagonal_State_Ensemble () + : nstates(0), state(Vect(1)), weight(0.0, 1) + { + } + + LiebLin_Diagonal_State_Ensemble::LiebLin_Diagonal_State_Ensemble (const LiebLin_Bethe_State& RefState, int nstates_req) + : nstates(nstates_req), state(Vect(RefState, nstates_req)), weight(1.0/nstates_req, nstates_req) + { + } + /* + LiebLin_Diagonal_State_Ensemble::LiebLin_Diagonal_State_Ensemble (const LiebLin_Bethe_State& RefState, int nstates_req, const Vect& weight_ref) + : nstates(nstates_req), state(Vect(RefState, nstates_req)), weight(weight_ref) + { + if (weight_ref.size() != nstates_req) JSCerror("Incompatible vector size in LiebLin_Diagonal_State_Ensemble constructor."); + } + */ + + // Recursively go through all arbitrary-order type 2 descendents + void Generate_type_2_descendents (LiebLin_Bethe_State& ActualState, int* ndesc_ptr, const LiebLin_Bethe_State& OriginState) + { + int type_required = 3; + Vect desc_label = Descendents (ActualState, OriginState, type_required); + + if (desc_label.size() > 0) { // follow down recursively + LiebLin_Bethe_State ActualState_here = ActualState; + for (int idesc = 0; idesc < desc_label.size(); ++idesc) { + ActualState_here.Set_to_Label (desc_label[idesc], OriginState.Ix2); + // output data: + cout << *ndesc_ptr << "\t" << desc_label[idesc] << endl; + cout << "origin: " << OriginState.Ix2 << endl; + cout << "shiftd: " << ActualState_here.Ix2 << endl; + ActualState_here.Compute_All(false); + cout << "\t\t\t\t\t\t" << ActualState.E - OriginState.E << endl; + *ndesc_ptr += 1; + Generate_type_2_descendents (ActualState_here, ndesc_ptr, OriginState); + + // do them mod st: + idesc += 6; + } + } + return; + } + + + + + //LiebLin_Diagonal_State_Ensemble::LiebLin_Diagonal_State_Ensemble (DP c_int, DP L, int N, const Root_Density& rho, int nstates_req) + //: nstates(nstates_req) + LiebLin_Diagonal_State_Ensemble::LiebLin_Diagonal_State_Ensemble (DP c_int, DP L, int N, const Root_Density& rho) + { + //version with 4 states in ensemble + + Root_Density x(rho.Npts, rho.lambdamax); + + for(int ix=0; ix Ix2_found(0.0, N); + int Ix2_left, Ix2_right; + Vect Ix2(N); + + + LiebLin_Bethe_State rhostate(c_int, N, L); + + nstates = 4; + state = Vect(rhostate, nstates); + weight = Vect(nstates); + weight[0] = 1.0/nstates; + weight[1] = 1.0/nstates; + weight[2] = 1.0/nstates; + weight[3] = 1.0/nstates; + + int n_moves = 0; + bool change = true; + + for (int i = 0; i < rho.Npts; ++i) { + integral_prev = integral; + integral += L * rho.value[i] * rho.dlambda[i]; + //cout << x.value[i] << "\t" << integral << endl; + if (integral > Nfound + 0.5) { + // Subtle error: if the rho is too discontinuous, i.e. if more than one rapidity is found, must correct for this. + if (integral > Nfound + 1.5 && integral < Nfound + 2.5) { // found two rapidities + JSCerror("The distribution of particles is too discontinous for discretisation"); + } + + else { + // few variables to clarify the computations, they do not need to be vectors, not used outside this loop + //Ix2_found[Nfound] = 2.0 * L * (x.value[i] * (integral - Nfound - 0.5) + x.value[i-1] * (Nfound + 0.5 - integral_prev))/(integral - integral_prev); + Ix2_found[Nfound] = 2.0 * L * x.Return_Value(rho.lambda[i]); + + Ix2_left = floor(Ix2_found[Nfound]); + Ix2_left -= (Ix2_left + N + 1)%2 ? 1 : 0; //adjust parity + + Ix2_right = ceil(Ix2_found[Nfound]); + Ix2_right += (Ix2_right + N + 1)%2 ? 1 : 0; //adjust parity + + int Ix2_in = (Ix2_found[Nfound] > 0 ? Ix2_left : Ix2_right); + int Ix2_out = (Ix2_found[Nfound] > 0 ? Ix2_right : Ix2_left); + cout << rho.lambda[i] << "\t" << x.Return_Value(rho.lambda[i]) << "\t" << Ix2_found[Nfound] << endl; + + //choose the saddle point state and remember the uncertain choices + if(Ix2_found[Nfound] - Ix2_left < 0.5) { + state[0].Ix2[Nfound] = Ix2_left; + state[1].Ix2[Nfound] = Ix2_left; + state[2].Ix2[Nfound] = Ix2_left; + state[3].Ix2[Nfound] = Ix2_left; + } + else if (Ix2_right - Ix2_found[Nfound] < 0.5) { + state[0].Ix2[Nfound] = Ix2_right; + state[1].Ix2[Nfound] = Ix2_right; + state[2].Ix2[Nfound] = Ix2_right; + state[3].Ix2[Nfound] = Ix2_right; + } + else { //it's a kind of magic: the zeroth goes in whle the first goes out, the second goes left if the third goes right + state[0].Ix2[Nfound] = Ix2_in; + state[1].Ix2[Nfound] = Ix2_out; + state[2+change].Ix2[Nfound] = Ix2_left; + state[2+!change].Ix2[Nfound] = Ix2_right; + change = !change; + ++n_moves; + } + Nfound++; + } + } + //cout << "\ti = " << i << "\tintegral = " << integral << "\tNfound = " << Nfound << endl; + } + //cout << endl; + + //fix state[3] and state[4] + for(int i=0; i Ix2_found(0.0, N); + int Ix2_left, Ix2_right; + Vect Ix2(N); + Vect Ix2_uncertain(0, N); + Vect index_uncertain(0, N); + int n_uncertain = 0, n_states_raw = 1; + + for (int i = 0; i < rho.Npts; ++i) { + integral_prev = integral; + integral += L * rho.value[i] * rho.dlambda[i]; + //cout << x.value[i] << "\t" << integral << endl; + if (integral > Nfound + 0.5) { + // Subtle error: if the rho is too discontinuous, i.e. if more than one rapidity is found, must correct for this. + if (integral > Nfound + 1.5 && integral < Nfound + 2.5) { // found two rapidities + JSCerror("The distribution of particles is too discontinous for discretisation"); + } + + else { + // few variables to clarify the computations, they do not need to be vectors, not used outside this loop + Ix2_found[Nfound] = 2.0 * L * (x.value[i] * (integral - Nfound - 0.5) + x.value[i-1] * (Nfound + 0.5 - integral_prev))/(integral - integral_prev); + + Ix2_left = floor(Ix2_found[Nfound]); + Ix2_left -= (Ix2_left + N + 1)%2 ? 1 : 0; //adjust parity + + Ix2_right = ceil(Ix2_found[Nfound]); + Ix2_right += (Ix2_right + N + 1)%2 ? 1 : 0; //adjust parity + + //cout << Ix2_found[Nfound] << "\t"; + + //choose the saddle point state and remember the uncertain choices + if(Ix2_found[Nfound] - Ix2_left < 1) { + Ix2[Nfound] = Ix2_left; + if(Ix2_found[Nfound] - Ix2_left > 0.75) { + Ix2_uncertain[n_uncertain] = -1; + index_uncertain[n_uncertain] = Nfound; + ++n_uncertain; + n_states_raw *= 2; + } + } + else if (Ix2_right - Ix2_found[Nfound] < 1) { + Ix2[Nfound] = Ix2_right; + if(Ix2_right - Ix2_found[Nfound] > 0.75) { + Ix2_uncertain[n_uncertain] = 1; + index_uncertain[n_uncertain] = Nfound; + ++n_uncertain; + n_states_raw *= 2; + } + } + else JSCerror("Cannot deduce a quantum number from x(lambda)"); + Nfound++; + } + } + //cout << "\ti = " << i << "\tintegral = " << integral << "\tNfound = " << Nfound << endl; + } + //cout << endl; + + //cout << Ix2 << endl; + + // create ensamble of states and compute its weights with respect to exact Ix2 saddle point + Vect > Ix2states(Ix2, n_states_raw); + Vect Ix2weight(0.0, n_states_raw); + +// Ix2states[0] = Ix2; +// Ix2weight[0] = ; + int n_states_proper = 0; + for(int i=0; i < n_states_raw; ++i) { + //only symmetric modifications + for(int mod_index = 0; mod_index> mod_index != 0) { + Ix2states[n_states_proper][index_uncertain[mod_index]] += (-2)*Ix2_uncertain[mod_index]; + ////Ix2states[state_index][N - 1 - hole_position[mod_index]] += (2)*holes[hole_position[mod_index]]; + } + } + //check if it's a proper set of quantum numbers + bool OK = true; + for(int j=0; j index(0, n_states_raw); + for (int nrs = 0; nrs < n_states_raw; ++nrs) index[nrs] = nrs; + Ix2weight.QuickSort(index); + + //cut the number of states if above 21 <- arbitrary number for now + nstates = min(n_states_proper, 21); + cout << n_states_raw << "\t" << n_states_proper << "\t" << nstates << endl; + + //create ensamble of states with normalised weights and ordered accordingly + LiebLin_Bethe_State rhostate(c_int, N, L); + rhostate.Ix2 = Ix2; + rhostate.Compute_All(true); + + state = Vect(rhostate, nstates); + weight = Vect(nstates); + + DP sum_weight = 0.0; + for(int i=0; i lambda_found(0.0, 2*N); + + for (int i = 0; i < rho.Npts; ++i) { + integral_prev = integral; + integral += L * rho.value[i] * rho.dlambda[i]; + //cout << x.value[i] << "\t" << integral << endl; + if (integral > Nfound + 0.5) { + cout << L*x.value[i] << "\t" << integral << endl; + // Subtle error: if the rho is too discontinuous, i.e. if more than one rapidity is found, must correct for this. + if (integral > Nfound + 1.5 && integral < Nfound + 2.5) { // found two rapidities + lambda_found[Nfound++] = 0.25 * (3.0 * rho.lambda[i-1] + rho.lambda[i]); + lambda_found[Nfound++] = 0.25 * (rho.lambda[i-1] + 3.0 * rho.lambda[i]); + } + else { + //lambda_found[Nfound] = rho.lambda[i]; + // Better: center the lambda_found between these points: + //lambda_found[Nfound] = rho.lambda[i-1] + (rho.lambda[i] - rho.lambda[i-1]) * ((Nfound + 1.0) - integral_prev)/(integral - integral_prev); + lambda_found[Nfound] = 0.5 * (rho.lambda[i-1] + rho.lambda[i]); + Nfound++; + } + } + //cout << "\ti = " << i << "\tintegral = " << integral << "\tNfound = " << Nfound << endl; + } + //cout << "rho: " << rho.Npts << " points" << endl << rho.value << endl; + //cout << "sym: " << rho.value[0] << " " << rho.value[rho.value.size() - 1] + // << "\t" << rho.value[rho.value.size()/2] << " " << rho.value[rho.value.size()/2 + 1] << endl; + //cout << "Found " << Nfound << " particles." << endl; + //cout << "lambda_found = " << lambda_found << endl; + + Vect lambda(N); + // Fill up the found rapidities: + for (int il = 0; il < JSC::min(N, Nfound); ++il) lambda[il] = lambda_found[il]; + // If there are missing ones, put them at the end; ideally, this should never be called + for (int il = Nfound; il < N; ++il) lambda[il] = lambda_found[Nfound-1] + (il - Nfound + 1) * (lambda_found[Nfound-1] - lambda_found[Nfound-2]); + + //cout << lambda << endl; +*/ + //try a different method + //first determine bins, that is intervals in lambda which contain a single rapidity +/* + Vect lambda_bin(0.0, N+1); + integral = 0.0; + Nfound = 0; + + lambda_bin[0] = 0; + lambda_bin[N] = rho.Npts-1; + for(int i=0; i Nfound + 1.0) lambda_bin[++Nfound] = i - (rho.lambda[i]>0); + } + + //put rapidity at the mean value of the distribution inside the bin + for(int ib=0; ib < N; ++ib) { + lambda_found[ib] = 0.0; + for(int i=lambda_bin[ib]; i Ix2_exact(N); + for (int i = 0; i < N; ++i) { + DP sum = 0.0; + for (int j = 0; j < N; ++j) sum += 2.0 * atan((lambda[i] - lambda[j])/c_int); + Ix2_exact[i] = (L * lambda[i] + sum)/PI; + Ix2_left[i] = floor(Ix2_exact[i]); + Ix2_left[i] -= (Ix2_left[i] + N + 1)%2 ? 1 : 0; + Ix2_right[i] = ceil(Ix2_exact[i]); + Ix2_right[i] += (Ix2_right[i] + N + 1)%2 ? 1 : 0; + + //cout << Ix2_left[i] << "\t" << Ix2_exact[i] << "\t" << Ix2_right[i] << endl; + //Ix2[i] = 2.0* int((L * lambda[i] + sum)/twoPI) + (N % 2) - 1; + // For N is even/odd, we want to round off to the nearest odd/even integer. + //Ix2[i] = 2.0 * floor((L* lambda[i] + sum)/twoPI + 0.5 * (N%2 ? 1 : 2)) + (N%2) - 1; + } + //cout << "Found quantum numbers " << endl << Ix2 << endl; +*/ + +/* + //create the reference state and mark possible hole positions + Vect Ix2(N); + Vect holes(0.0, N); + for(int i=0; i < N; ++i) { + if(Ix2_found[i] - Ix2_left[i] < 1) { + Ix2[i] = Ix2_left[i]; + if(Ix2_found[i] - Ix2_left[i] > 0.75) holes[i] = -1; + } + else { + Ix2[i] = Ix2_right[i]; + if(Ix2_right[i] - Ix2_found[i] > 0.75) holes[i] = 1; + } + } +*/ +// cout << endl << Ix2 << endl; +// cout << endl << holes << endl; +/* + //count number of possible states + int n = 0; + for(int i=0; i < N; ++i) { + if(holes[i] != 0) ++n; + } + int n_states_raw = 1; +*/ +/* only symmetric modifications + for(int i=0; i<0.5*n; ++i) n_states_raw *= 2.0; //we count only symmetric modiications + + Vect > Ix2states(Ix2, n_states_raw); + + Vect hole_position(0, 0.5*n); + int hole_index = 0; + for(int i=N/2 + N%2; i> mod_index != 0) { + Ix2states[state_index][hole_position[mod_index]] += (-2)*holes[hole_position[mod_index]]; + Ix2states[state_index][N - 1 - hole_position[mod_index]] += (2)*holes[hole_position[mod_index]]; + } + } + //check if it's a proper set of quantum numbers + bool OK = true; + for(int j=0; j > Ix2states(Ix2, n_states_raw); + + Vect hole_position(0, n); + int hole_index = 0; + for(int i=0; i> mod_index != 0) { + Ix2states[state_index][hole_position[mod_index]] += (-2)*holes[hole_position[mod_index]]; + //Ix2states[state_index][N - 1 - hole_position[mod_index]] += (2)*holes[hole_position[mod_index]]; + } + } + //check if it's a proper set of quantum numbers + bool OK = true; + for(int j=0; j state_raw = Vect(rhostate, n_states_raw); + Vect weight_raw(0.0, n_states_raw); +*/ + +/* + DP energy_rho = 0; + for (int i = 0; i < rho.Npts; ++i) { + energy_rho += L * rho.value[i] * rho.dlambda[i] * rho.lambda[i] * rho.lambda[i]; + } + + cout << energy_rho << endl; + + DP energy_discrete = 0; + for (int i=0; i < N; ++i) { + energy_discrete += lambda[i] * lambda[i]; + } + energy_discrete /= N; + cout << energy_discrete; +*/ +/* + DP sum_weight_raw = 0.0; + for(int i=0; i index(n_states_raw); + for (int nrs = 0; nrs < n_states_raw; ++nrs) index[nrs] = nrs; + weight_raw.QuickSort(index); + + //nstates = 0; + //for(int i=0; i 0.01) ++nstates; + + nstates = JSC::min(n_states_raw, 21); + + cout << nstates << endl; + + state = Vect(rhostate, nstates); + weight = Vect(nstates); + + DP sum_weight = 0.0; + for(int i=0; i 0) Ix2[i] += 2; + allOK = true; + for (int i = 0; i < N-1; ++i) if (Ix2[i] == Ix2[i+1]) allOK = false; + } + //cout << "Found modified quantum numbers " << endl << Ix2 << endl; + + LiebLin_Bethe_State rhostate(c_int, N, L); + rhostate.Ix2 = Ix2; + rhostate.Compute_All(true); + //cout << "rapidities of state found: " << rhostate.lambda << endl; + + // Until here, the code is identical to that in Discretized_LiebLin_Bethe_State. + // Construct states in the vicinity by going through type 2 descendents recursively: + int ndesc_type2 = 0; + int* ndesc_type2_ptr = &ndesc_type2; + Generate_type_2_descendents (rhostate, ndesc_type2_ptr, rhostate); + + + cout << "Found " << ndesc_type2 << " descendents for state " << rhostate << endl; + //JSCerror("Stop here..."); + + // Now try to construct states in the vicinity. + int nrstates1mod = 1; + for (int i = 0; i < N; ++i) { + if (i == 0 || Ix2[i-1] < Ix2[i] - 2) nrstates1mod++; + if (i == N-1 || Ix2[i+1] > Ix2[i] + 2) nrstates1mod++; + } + //if (nrstates1mod < nstates_req) JSCerror("nrstates1mod < nstates_req in LiebLin_Diagonal_State_Ensemble."); + nstates = nrstates1mod; + + Vect > Ix2states1mod(nrstates1mod); + for (int nrs = 0; nrs < nrstates1mod; ++nrs) Ix2states1mod[nrs] = Vect (N); + + int nfound = 0; + Ix2states1mod[nfound++] = Ix2; // this is the sp state + for (int i = 0; i < N; ++i) { + if (i == 0 || Ix2[i-1] < Ix2[i] - 2) { + Ix2states1mod[nfound] = Ix2; + Ix2states1mod[nfound++][i] -= 2; + } + if (i == N-1 || Ix2[i+1] > Ix2[i] + 2) { + Ix2states1mod[nfound] = Ix2; + Ix2states1mod[nfound++][i] += 2; + } + } + + // Evaluate the weight of all found states: + Vect rawweight(nrstates1mod); + + for (int nrs = 0; nrs < nrstates1mod; ++nrs) { + DP sumweight = 0.0; + for (int i = 0; i < N; ++i) sumweight += fabs(Ix2states1mod[nrs][i] - Ix2_exact[i]); + rawweight[nrs] = exp(-sumweight/log(L)); + } + + // Order the weights in decreasing value: + Vect index(nrstates1mod); + for (int nrs = 0; nrs < nrstates1mod; ++nrs) index[nrs] = nrs; + rawweight.QuickSort(index); + + // Calculate weight normalization: + DP weightnorm = 0.0; + //for (int ns = 0; ns < nstates_req; ++ns) weightnorm += rawweight[nrstates1mod - 1 - ns]; + for (int ns = 0; ns < nstates; ++ns) weightnorm += rawweight[nrstates1mod - 1 - ns]; + + state = Vect(rhostate, nstates); + weight = Vect (nstates); + + for (int ns = 0; ns < nstates; ++ns) { + weight[ns] = rawweight[nrstates1mod - 1 - ns]/weightnorm; + state[ns].Ix2 = Ix2states1mod[index[nrstates1mod - 1 - ns] ]; + state[ns].Set_Label_from_Ix2 (rhostate.Ix2); + state[ns].Compute_All(true); + } +*/ + + + } + + /* + LiebLin_Diagonal_State_Ensemble::LiebLin_Diagonal_State_Ensemble (DP c_int, DP L, int N, const Root_Density& rho, int nstates_req) + : nstates(nstates_req) + { + // This function returns a state ensemble matching the continuous density rho. + // The logic closely resembles the one used in Discretized_LiebLin_Bethe_State. + + // Each `exact' (from rho) quantum number is matched to its two possible values, + // assigning a probability to each according to the proximity of the quantum number values. + + // The set is then ordered in energy, and split into nstates_req boxes from which one + // representative is selected, and given the probability = sum of probabilities in the box. + + + // Begin as per Discretized_LiebLin_Bethe_State, to find the saddle-point state. + // Each time N \int_{-\infty}^\lambda d\lambda' \rho(\lambda') crosses a half integer, add a particle: + DP integral = 0.0; + DP integral_prev = 0.0; + int Nfound = 0; + Vect lambda_found(0.0, 2*N); + + for (int i = 0; i < rho.Npts; ++i) { + integral_prev = integral; + integral += L * rho.value[i] * rho.dlambda[i]; + if (integral > Nfound + 0.5) { + // Subtle error: if the rho is too discontinuous, i.e. if more than one rapidity is found, must correct for this. + if (integral > Nfound + 1.5 && integral < Nfound + 2.5) { // found two rapidities + lambda_found[Nfound++] = 0.25 * (3.0 * rho.lambda[i-1] + rho.lambda[i]); + lambda_found[Nfound++] = 0.25 * (rho.lambda[i-1] + 3.0 * rho.lambda[i]); + } + else { + //lambda_found[Nfound] = rho.lambda[i]; + // Better: center the lambda_found between these points: + //lambda_found[Nfound] = rho.lambda[i-1] + (rho.lambda[i] - rho.lambda[i-1]) * ((Nfound + 1.0) - integral_prev)/(integral - integral_prev); + lambda_found[Nfound] = 0.5 * (rho.lambda[i-1] + rho.lambda[i]); + Nfound++; + } + } + //cout << "\ti = " << i << "\tintegral = " << integral << "\tNfound = " << Nfound << endl; + } + //cout << "rho: " << rho.Npts << " points" << endl << rho.value << endl; + //cout << "sym: " << rho.value[0] << " " << rho.value[rho.value.size() - 1] + // << "\t" << rho.value[rho.value.size()/2] << " " << rho.value[rho.value.size()/2 + 1] << endl; + //cout << "Found " << Nfound << " particles." << endl; + //cout << "lambda_found = " << lambda_found << endl; + + Vect lambda(N); + // Fill up the found rapidities: + for (int il = 0; il < JSC::min(N, Nfound); ++il) lambda[il] = lambda_found[il]; + // If there are missing ones, put them at the end; ideally, this should never be called + for (int il = Nfound; il < N; ++il) lambda[il] = lambda_found[Nfound-1] + (il - Nfound + 1) * (lambda_found[Nfound-1] - lambda_found[Nfound-2]); + + // Calculate quantum numbers: 2\pi * (L lambda + \sum_j 2 atan((lambda - lambda_j)/c)) = I_j + // Use rounding. + Vect Ix2_exact(N); + Vect Ix2(N); + for (int i = 0; i < N; ++i) { + DP sum = 0.0; + for (int j = 0; j < N; ++j) sum += 2.0 * atan((lambda[i] - lambda[j])/c_int); + Ix2_exact[i] = (L * lambda[i] + sum)/PI; + //Ix2[i] = 2.0* int((L * lambda[i] + sum)/twoPI) + (N % 2) - 1; + // For N is even/odd, we want to round off to the nearest odd/even integer. + Ix2[i] = 2.0 * floor((L* lambda[i] + sum)/twoPI + 0.5 * (N%2 ? 1 : 2)) + (N%2) - 1; + } + //cout << "Found quantum numbers " << endl << Ix2 << endl; + + // Check that the quantum numbers are all distinct: + bool allOK = false; + while (!allOK) { + for (int i = 0; i < N-1; ++i) if (Ix2[i] == Ix2[i+1] && Ix2[i] < 0) Ix2[i] -= 2; + for (int i = 1; i < N; ++i) if (Ix2[i] == Ix2[i-1] && Ix2[i] > 0) Ix2[i] += 2; + allOK = true; + for (int i = 0; i < N-1; ++i) if (Ix2[i] == Ix2[i+1]) allOK = false; + } + //cout << "Found modified quantum numbers " << endl << Ix2 << endl; + + LiebLin_Bethe_State rhostate(c_int, N, L); + rhostate.Ix2 = Ix2; + rhostate.Compute_All(true); + //cout << "rapidities of state found: " << rhostate.lambda << endl; + + // Until here, the code is identical to that in Discretized_LiebLin_Bethe_State. + + // Now try to construct states in the vicinity. + int nrstates1mod = 1; + for (int i = 0; i < N; ++i) { + if (i == 0 || Ix2[i-1] < Ix2[i] - 2) nrstates1mod++; + if (i == N-1 || Ix2[i+1] > Ix2[i] + 2) nrstates1mod++; + } + if (nrstates1mod < nstates_req) JSCerror("nrstates1mod < nstates_req in LiebLin_Diagonal_State_Ensemble."); + + Vect > Ix2states1mod(nrstates1mod); + for (int nrs = 0; nrs < nrstates1mod; ++nrs) Ix2states1mod[nrs] = Vect (N); + + int nfound = 0; + Ix2states1mod[nfound++] = Ix2; // this is the sp state + for (int i = 0; i < N; ++i) { + if (i == 0 || Ix2[i-1] < Ix2[i] - 2) { + Ix2states1mod[nfound] = Ix2; + Ix2states1mod[nfound++][i] -= 2; + } + if (i == N-1 || Ix2[i+1] > Ix2[i] + 2) { + Ix2states1mod[nfound] = Ix2; + Ix2states1mod[nfound++][i] += 2; + } + } + + // Evaluate the weight of all found states: + Vect rawweight(nrstates1mod); + + for (int nrs = 0; nrs < nrstates1mod; ++nrs) { + DP sumweight = 0.0; + for (int i = 0; i < N; ++i) sumweight += fabs(Ix2states1mod[nrs][i] - Ix2_exact[i]); + rawweight[nrs] = exp(-sumweight/log(L)); + } + + // Order the weights in decreasing value: + Vect index(nrstates1mod); + for (int nrs = 0; nrs < nrstates1mod; ++nrs) index[nrs] = nrs; + rawweight.QuickSort(index); + + // Calculate weight normalization: + DP weightnorm = 0.0; + for (int ns = 0; ns < nstates_req; ++ns) weightnorm += rawweight[nrstates1mod - 1 - ns]; + + state = Vect(rhostate, nstates_req); + weight = Vect (nstates_req); + + for (int ns = 0; ns < nstates_req; ++ns) { + weight[ns] = rawweight[nrstates1mod - 1 - ns]/weightnorm; + state[ns].Ix2 = Ix2states1mod[index[nrstates1mod - 1 - ns] ]; + state[ns].Set_Label_from_Ix2 (rhostate.Ix2); + state[ns].Compute_All(true); + } + } + */ + + LiebLin_Diagonal_State_Ensemble& LiebLin_Diagonal_State_Ensemble::operator= (const LiebLin_Diagonal_State_Ensemble& rhs) + { + if (this != &rhs) { + nstates = rhs.nstates; + state = rhs.state; + weight = rhs.weight; + } + + return(*this); + } + + void LiebLin_Diagonal_State_Ensemble::Load (DP c_int, DP L, int N, const char* ensfile_Cstr) + { + ifstream infile(ensfile_Cstr); + + // Count nr of lines in file: + string dummy; + int nrlines = 0; + while (getline(infile, dummy)) ++nrlines; + infile.close(); + + //cout << "Found " << nrlines << " lines in ens file." << endl; + nstates = nrlines; + LiebLin_Bethe_State examplestate (c_int, L, N); + state = Vect (examplestate, nstates); + weight = Vect (0.0, nstates); + + infile.open(ensfile_Cstr); + + string dummylabel; + for (int ns = 0; ns < nstates; ++ns) { + infile >> weight[ns] >> dummylabel; + for (int i = 0; i < state[ns].N; ++i) infile >> state[ns].Ix2[i]; + } + infile.close(); + + for (int ns = 0; ns < nstates; ++ns) { + state[ns].Set_Label_from_Ix2 (state[0].Ix2); // labels are ref to the saddle-point state + state[ns].Compute_All(true); + } + } + + void LiebLin_Diagonal_State_Ensemble::Save (const char* ensfile_Cstr) + { + ofstream outfile; + + outfile.open(ensfile_Cstr); + + for (int ns = 0; ns < nstates; ++ns) { + if (ns > 0) outfile << endl; + outfile << setprecision(16) << weight[ns] << "\t" << state[ns].label << "\t"; + for (int i = 0; i < state[ns].N; ++i) outfile << " " << state[ns].Ix2[i]; + } + } + + + //LiebLin_Diagonal_State_Ensemble LiebLin_Thermal_Saddle_Point_Ensemble (DP c_int, DP L, int N, DP kBT, int nstates_req) + LiebLin_Diagonal_State_Ensemble LiebLin_Thermal_Saddle_Point_Ensemble (DP c_int, DP L, int N, DP kBT) + { + // This function constructs a manifold of states around the thermal saddle-point. + + // Start as per Canonical_Saddle_Point_State: + + // This function returns the discretized state minimizing the canonical free energy + // F = E - T S. + + // This is obtained by rediscretizing the solution coming from TBA. + + + // Otherwise, return the discretized TBA saddle-point state: + LiebLin_TBA_Solution TBAsol = LiebLin_TBA_Solution_fixed_nbar (c_int, N/L, kBT, 1.0e-4, 100); + + LiebLin_Diagonal_State_Ensemble ensemble(c_int, L, N, TBAsol.rho); + + cout << "nbar: " << TBAsol.nbar << endl; + cout << "ebar: " << TBAsol.ebar << endl; + + cout << ensemble.state[0].E << endl; + //return(LiebLin_Diagonal_State_Ensemble (c_int, L, N, TBAsol.rho, nstates_req)); + return(ensemble); + } + +} // namespace JSC + + + + + + diff --git a/src/LIEBLIN/LiebLin_Sumrules.cc b/src/LIEBLIN/LiebLin_Sumrules.cc new file mode 100644 index 0000000..da849d3 --- /dev/null +++ b/src/LIEBLIN/LiebLin_Sumrules.cc @@ -0,0 +1,302 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/LIEBLIN/LiebLin_Sumrules.cc + +Purpose: provides functions evaluating various sumrule factors + for Lieb-Liniger. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + //DP Sumrule_Factor (char whichDSF, LiebLin_Bethe_State& RefState, DP Chem_Pot, bool fixed_iK, int iKneeded) + DP Sumrule_Factor (char whichDSF, LiebLin_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax) + { + DP sumrule_factor = 1.0; + + //if (!fixed_iK) { + if (iKmin != iKmax) { + if (whichDSF == 'Z') sumrule_factor = 1.0; + else if (whichDSF == 'd' || whichDSF == '1') { + + // Here, we use a measure decreasing in K with K^2. + // We sum up omega * MEsq/(iK^2) for all values of iKmin <= iK <= iKmax, discounting iK == 0 (where DSF vanishes) + // We therefore have (N/L) x L^{-1} x (2\pi/L)^2 x (iKmax - iKmin + 1) = 4 \pi^2 x N x (iKmax - iKmin + 1)/L^4 + // Discounting iK == 0 (where DSF vanishes), if iKmin <= 0 && iKmax >= 0 (in which case 0 is containted in [iKmin, iKmax]) + sumrule_factor = (iKmin <= 0 && iKmax >= 0) ? + (RefState.L * RefState.L * RefState.L * RefState.L)/(4.0 * PI * PI * RefState.N * (iKmax - iKmin)) + : (RefState.L * RefState.L * RefState.L * RefState.L)/(4.0 * PI * PI * RefState.N * (iKmax - iKmin + 1)); + + /* + // Measure using the g2(0) + delta function: // DOES NOT WORK VERY WELL + DP dE0_dc = LiebLin_dE0_dc (RefState.c_int, RefState.L, RefState.N); + //sumrule_factor = 1.0/((dE0_dc + (2.0 * RefState.Tableau[0].Ncols + 1.0)*RefState.N/RefState.L)/RefState.L); + // Assume that iKmin == 0 here: + //sumrule_factor = 1.0/((dE0_dc + (2*iKmax + 1)*RefState.N/RefState.L)/RefState.L); + // For iKmin != 0: + sumrule_factor = 1.0/((dE0_dc + (iKmax - iKmin + 1)*RefState.N/RefState.L)/RefState.L); + */ + } + // For the Green's function, it's the delta function \delta(x = 0) plus the density: + //else if (whichDSF == 'g') sumrule_factor = 1.0/((2.0 * RefState.Tableau[0].Ncols + 1.0)/RefState.L + RefState.N/RefState.L); + // Assume that iKmin == 0 here: + //else if (whichDSF == 'g') sumrule_factor = 1.0/(2.0* iKmax + 1.0)/RefState.L + RefState.N/RefState.L); + else if (whichDSF == 'g') + sumrule_factor = 1.0/((abs(iKmax - iKmin) + 1.0)/RefState.L + RefState.N/RefState.L); + //sumrule_factor = 1.0/((pow(twoPI * iKmax/RefState.L, 2.0) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L)/RefState.L); + // For the one-body function, it's just the density: + else if (whichDSF == 'o') sumrule_factor = RefState.L/RefState.N; + else if (whichDSF == 'q') sumrule_factor = 1.0; + else if (whichDSF == 'B') sumrule_factor = 1.0; + else if (whichDSF == 'C') sumrule_factor = 1.0; + else JSCerror("whichDSF option not consistent in Sumrule_Factor"); + } + + //else if (fixed_iK) { + else if (iKmin == iKmax) { + if (whichDSF == 'Z') sumrule_factor = 1.0; + else if (whichDSF == 'd' || whichDSF == '1') + //// We sum up omega * MEsq/(iK^2): this should give (1/L) x (N/L) x k^2 = N x (2\pi)^2/L^4 + //sumrule_factor = pow(RefState.L, 4.0)/(4.0 * PI * PI * RefState.N); + // We sum up omega * MEsq + //sumrule_factor = pow(RefState.L, 4.0)/(4.0 * PI * PI * iKneeded * iKneeded * RefState.N); + sumrule_factor = pow(RefState.L, 4.0)/(4.0 * PI * PI * iKmax * iKmax * RefState.N); + else if (whichDSF == 'g' || whichDSF == 'o') { + // We sum up omega * MEsq + //sumrule_factor = 1.0/((pow(twoPI * iKneeded/RefState.L, 2.0) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L)/RefState.L); + sumrule_factor = 1.0/((pow(twoPI * iKmax/RefState.L, 2.0) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L)/RefState.L); + } + //else if (whichDSF == 'o') sumrule_factor = RefState.L/RefState.N; + else if (whichDSF == 'q') sumrule_factor = 1.0; + else if (whichDSF == 'B') sumrule_factor = 1.0; + else if (whichDSF == 'C') sumrule_factor = 1.0; + else JSCerror("whichDSF option not consistent in Sumrule_Factor"); + } + + return(sumrule_factor); + } + + void Evaluate_F_Sumrule (char whichDSF, const LiebLin_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax, const char* RAW_Cstr, const char* FSR_Cstr) + { + + ifstream infile; + infile.open(RAW_Cstr); + if(infile.fail()) { + cout << "Filename RAW_Cstr = " << RAW_Cstr << endl; + JSCerror("Could not open input file in Evaluate_F_Sumrule(LiebLin...)."); + } + + // We run through the data file to check the f sumrule at each positive momenta: + //int iK_UL = RefState.Tableau[0].Ncols; // this is iK_UL + //Vect Sum_omega_MEsq(0.0, iK_UL + 1); + Vect_DP Sum_omega_MEsq (0.0, iKmax - iKmin + 1); + Vect_DP Sum_abs_omega_MEsq (0.0, iKmax - iKmin + 1); + + DP Sum_MEsq = 0.0; + + DP omega, ME; + int iK; + //int conv; + DP dev; + string label; + int nr, nl; + + int nraw = 0; + + while (infile.peek() != EOF) { + nraw++; + infile >> omega >> iK >> ME >> dev >> label; + if (whichDSF == '1') infile >> nr >> nl; + //if (iK > 0 && iK <= iK_UL) Sum_omega_MEsq[iK] += omega * MEsq; + if (iK >= iKmin && iK <= iKmax) Sum_omega_MEsq[iK - iKmin] += omega * ME * ME; + if (iK >= iKmin && iK <= iKmax) Sum_abs_omega_MEsq[iK - iKmin] += fabs(omega * ME * ME); + Sum_MEsq += ME * ME; + } + + infile.close(); + + //cout << "Read " << nraw << " entries in raw file." << endl; + + ofstream outfile; + outfile.open(FSR_Cstr); + outfile.precision(16); + + if (whichDSF == 'd' || whichDSF == '1') { + /* + outfile << 0 << "\t" << 1; // full saturation at k = 0 ! + for (int i = 1; i <= iK_UL; ++i) + outfile << endl << i << "\t" << Sum_omega_MEsq[i] * RefState.L * RefState.L + * RefState.L * RefState.L/(4.0 * PI * PI * i * i * RefState.N); + */ + + for (int i = iKmin; i <= iKmax; ++i) { + + if (i > iKmin) outfile << endl; + //outfile << i << "\t" << Sum_omega_MEsq[i - iKmin] * pow(RefState.L, 4.0)/(pow(2.0 * PI * JSC::max(abs(i), 1), 2.0) * RefState.N); + outfile << i << "\t" << Sum_omega_MEsq[i - iKmin] * pow(RefState.L, 4.0)/(pow(2.0 * PI * JSC::max(abs(i), 1), 2.0) * RefState.N) + // Include average of result at +iK and -iK in a third column: iK is at index index(iK) = iK - iKmin + // so -iK is at index index(-iK) = -iK - iKmin + // We can only use this index if it is >= 0 and < iKmax - iKmin + 1, otherwise third column is copy of second: + << "\t" << ((i + iKmin <= 0 && -i < iKmax + 1) ? + 0.5 * (Sum_omega_MEsq[i - iKmin] + Sum_omega_MEsq[-i - iKmin]) + : Sum_omega_MEsq[i - iKmin]) + * pow(RefState.L, 4.0)/(pow(2.0 * PI * JSC::max(abs(i), 1), 2.0) * RefState.N); + } + } + else if (whichDSF == 'g' || whichDSF == 'o') { + /* + for (int i = 0; i <= iK_UL; ++i) + outfile << endl << i << "\t" << Sum_omega_MEsq[i] * RefState.L + /((4.0 * PI * PI * i * i)/(RefState.L * RefState.L) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L); + //cout << "Sum_MEsq = " << Sum_MEsq << "\tN/L = " << RefState.N/RefState.L << endl; + */ + for (int i = iKmin; i <= iKmax; ++i) { + if (i > iKmin) outfile << endl; + //outfile << i << "\t" << Sum_omega_MEsq[i - iKmin] * RefState.L + ///((4.0 * PI * PI * i * i)/(RefState.L * RefState.L) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L); + outfile << i << "\t" << Sum_omega_MEsq[i - iKmin] * RefState.L + /((4.0 * PI * PI * i * i)/(RefState.L * RefState.L) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L) + << "\t" << ((i + iKmin <= 0 && -i < iKmax + 1) ? + 0.5 * (Sum_omega_MEsq[i - iKmin] + Sum_omega_MEsq[-i - iKmin]) : Sum_omega_MEsq[i - iKmin]) + * RefState.L/((4.0 * PI * PI * i * i)/(RefState.L * RefState.L) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L); + } + } + + outfile.close(); + + } + + void Evaluate_F_Sumrule (string prefix, char whichDSF, const LiebLin_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax) + { + + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + stringstream FSR_stringstream; string FSR_string; + FSR_stringstream << prefix << ".fsr"; + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + + Evaluate_F_Sumrule (whichDSF, RefState, Chem_Pot, iKmin, iKmax, RAW_Cstr, FSR_Cstr); + + } + + // Using diagonal state ensemble: + void Evaluate_F_Sumrule (char whichDSF, DP c_int, DP L, int N, DP kBT, int nstates_req, DP Chem_Pot, int iKmin, int iKmax, const char* FSR_Cstr) + { + + // We run through the data file to check the f sumrule at each positive momenta: + Vect_DP Sum_omega_MEsq (0.0, iKmax - iKmin + 1); + + DP Sum_MEsq = 0.0; + + DP omega, ME; + int iK; + //int conv; + DP dev; + string label; + int nr, nl; + + // Read the weights from the ensembles file: + LiebLin_Diagonal_State_Ensemble ensemble; + + stringstream ensfilestrstream; + //ensfilestrstream << "LiebLin_c_int_" << c_int << "_L_" << L << "_N_" << N << "_kBT_" << kBT << "_ns_" << nstates_req << ".ens"; + ensfilestrstream << "LiebLin_c_int_" << c_int << "_L_" << L << "_N_" << N << "_kBT_" << kBT << ".ens"; + string ensfilestr = ensfilestrstream.str(); + const char* ensfile_Cstr = ensfilestr.c_str(); + + ensemble.Load(c_int, L, N, ensfile_Cstr); + + for (int ns = 0; ns < ensemble.nstates; ++ns) { + + // Define the raw input file name: + stringstream filenameprefix; + //Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, kBT, ensemble.state[ns], ensemble.state[ns], ensemble.state[ns].label); + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, 0.0, ensemble.state[ns], ensemble.state[ns], ensemble.state[ns].label); + string prefix = filenameprefix.str(); + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + ifstream infile; + infile.open(RAW_Cstr); + if(infile.fail()) { + cout << "Filename RAW_Cstr = " << RAW_Cstr << endl; + JSCerror("Could not open input file in Evaluate_F_Sumrule(LiebLin...)."); + } + + while (infile.peek() != EOF) { + infile >> omega >> iK >> ME >> dev >> label; + if (whichDSF == '1') infile >> nr >> nl; + //if (iK > 0 && iK <= iK_UL) Sum_omega_MEsq[iK] += omega * MEsq; + if (iK >= iKmin && iK <= iKmax) Sum_omega_MEsq[iK - iKmin] += ensemble.weight[ns] * omega * ME * ME; + Sum_MEsq += ensemble.weight[ns] * ME * ME; + } + + infile.close(); + } + + LiebLin_Bethe_State RefState = ensemble.state[0]; // to use the code below, which comes from earlier Evaluate_F_Sumrule + + ofstream outfile; + outfile.open(FSR_Cstr); + outfile.precision(16); + + if (whichDSF == 'd' || whichDSF == '1') { + /* + outfile << 0 << "\t" << 1; // full saturation at k = 0 ! + for (int i = 1; i <= iK_UL; ++i) + outfile << endl << i << "\t" << Sum_omega_MEsq[i] * RefState.L * RefState.L + * RefState.L * RefState.L/(4.0 * PI * PI * i * i * RefState.N); + */ + for (int i = iKmin; i <= iKmax; ++i) { + if (i > iKmin) outfile << endl; + //outfile << i << "\t" << Sum_omega_MEsq[i - iKmin] * pow(RefState.L, 4.0)/(pow(2.0 * PI * JSC::max(abs(i), 1), 2.0) * RefState.N); + outfile << i << "\t" << Sum_omega_MEsq[i - iKmin] * pow(RefState.L, 4.0)/(pow(2.0 * PI * JSC::max(abs(i), 1), 2.0) * RefState.N) + // Include average of result at +iK and -iK in a third column: iK is at index index(iK) = iK - iKmin + // so -iK is at index index(-iK) = -iK - iKmin + // We can only use this index if it is >= 0 and < iKmax - iKmin + 1, otherwise third column is copy of second: + << "\t" << ((i + iKmin <= 0 && -i < iKmax + 1) ? + 0.5 * (Sum_omega_MEsq[i - iKmin] + Sum_omega_MEsq[-i - iKmin]) + : Sum_omega_MEsq[i - iKmin]) + * pow(RefState.L, 4.0)/(pow(2.0 * PI * JSC::max(abs(i), 1), 2.0) * RefState.N); + } + } + else if (whichDSF == 'g' || whichDSF == 'o') { + /* + for (int i = 0; i <= iK_UL; ++i) + outfile << endl << i << "\t" << Sum_omega_MEsq[i] * RefState.L + /((4.0 * PI * PI * i * i)/(RefState.L * RefState.L) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L); + //cout << "Sum_MEsq = " << Sum_MEsq << "\tN/L = " << RefState.N/RefState.L << endl; + */ + for (int i = iKmin; i <= iKmax; ++i) { + if (i > iKmin) outfile << endl; + //outfile << i << "\t" << Sum_omega_MEsq[i - iKmin] * RefState.L + ///((4.0 * PI * PI * i * i)/(RefState.L * RefState.L) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L); + outfile << i << "\t" << Sum_omega_MEsq[i - iKmin] * RefState.L + /((4.0 * PI * PI * i * i)/(RefState.L * RefState.L) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L) + << "\t" << ((i + iKmin <= 0 && -i < iKmax + 1) ? + 0.5 * (Sum_omega_MEsq[i - iKmin] + Sum_omega_MEsq[-i - iKmin]) : Sum_omega_MEsq[i - iKmin]) + * RefState.L/((4.0 * PI * PI * i * i)/(RefState.L * RefState.L) - Chem_Pot + 4.0 * RefState.c_int * RefState.N/RefState.L); + } + } + + outfile.close(); + + } + + +} // namespace JSC diff --git a/src/LIEBLIN/LiebLin_Tgt0.cc b/src/LIEBLIN/LiebLin_Tgt0.cc new file mode 100644 index 0000000..f0ad8ae --- /dev/null +++ b/src/LIEBLIN/LiebLin_Tgt0.cc @@ -0,0 +1,536 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Tgt0.cc + +Purpose: Finite temperature correlations for Lieb-Liniger + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + + /* + DP Entropy (LiebLin_Bethe_State& RefState, DP epsilon) + { + // This function calculates the entropy of a finite Lieb-Liniger state, + // using the quantum number space particle and hole densities. + + // The densities are a sum of Gaussians with width epsilon. + + // We calculate \rho (x) and \rho_h (x) for x_j on a fine lattice. + + int Ix2max = RefState.Ix2.max(); + int Ix2min = RefState.Ix2.min(); + + //cout << Ix2max << "\t" << Ix2min << endl; + + // Give a bit of room for leftmost and rightmost particles Gaussians to attain asymptotic values: + Ix2max += int(10.0 * RefState.L * epsilon); + Ix2min -= int(10.0 * RefState.L * epsilon); + + //cout << Ix2max << "\t" << Ix2min << endl; + + DP xmax = Ix2max/(2.0* RefState.L); + DP xmin = Ix2min/(2.0* RefState.L); + + //xmax += 10.0* epsilon; + //xmin -= 10.0* epsilon; + + DP dx = 0.1/RefState.L; + int Nptsx = int((xmax - xmin)/dx); + + Vect occupied (false, (Ix2max - Ix2min)/2 + 1); // whether there is a particle or not + for (int i = 0; i < RefState.N; ++i) occupied[(RefState.Ix2[i] - Ix2min)/2] = true; + + Vect rho(0.0, Nptsx); + Vect rhoh(0.0, Nptsx); + + //cout << xmin << "\t" << xmax << "\t" << dx << "\t" << Nptsx << endl; + + DP x; + DP epsilonsq = epsilon * epsilon; + DP twoepsilonsq = 2.0 * epsilon * epsilon; + + //Vect xparticle(RefState.N); + //for (int i = 0; i < RefState.N; ++i) xparticle[i] = RefState.Ix2[i]/(2.0* RefState.L); + Vect xarray((Ix2max - Ix2min)/2 + 1); + for (int i = 0; i < (Ix2max - Ix2min)/2 + 1; ++i) xarray[i] = (0.5*Ix2min + i)/RefState.L; + + for (int ix = 0; ix < Nptsx; ++ix) { + x = xmin + dx * (ix + 0.5); + //for (int i = 0; i < RefState.N; ++i) rho[ix] += 1.0/((x - xparticle[i]) * (x - xparticle[i]) + epsilonsq); + //rho[ix] *= epsilon/(PI * RefState.L); + for (int i = 0; i < (Ix2max - Ix2min)/2 + 1; ++i) { + // Using Gaussians: + //if (occupied[i]) rho[ix] += exp(-(x - xarray[i]) * (x - xarray[i])/twoepsilonsq); + //else rhoh[ix] += exp(-(x - xarray[i]) * (x - xarray[i])/twoepsilonsq); + // Using Lorentzians: + if (occupied[i]) rho[ix] += 1.0/((x - xarray[i]) * (x - xarray[i]) + epsilonsq); + else rhoh[ix] += 1.0/((x - xarray[i]) * (x - xarray[i]) + epsilonsq); + } + //rho[ix] /= sqrt(twoPI) * epsilon * RefState.L; + //rhoh[ix] /= sqrt(twoPI) * epsilon * RefState.L; + rho[ix] *= epsilon/(PI * RefState.L); + rhoh[ix] *= epsilon/(PI * RefState.L); + //cout << ix << " x = " << x << "\trho = " << rho[ix] << "\trhoh = " << rhoh[ix] << "\trho + rhoh = " << rho[ix] + rhoh[ix] << endl; + } + + // Now calculate the entropy: + complex entropy = 0.0; + DP Deltax = log(RefState.L)/RefState.L; + for (int ix = 0; ix < Nptsx; ++ix) + //entropy -= ln_Gamma (1.0 + rho[ix]) + ln_Gamma (2.0 - rho[ix]); // This is ln (\rho_tot choose \rho) with \rho_tot = 1. + entropy += ln_Gamma (RefState.L * (rho[ix] + rhoh[ix]) * Deltax + 1.0) - ln_Gamma(RefState.L * rho[ix] * Deltax + 1.0) - ln_Gamma (RefState.L * rhoh[ix] * Deltax + 1.0); // This is ln (\rho_tot choose \rho) with \rho_tot = 1. + entropy *= dx/Deltax; + + //cout << "Entropy found " << entropy << "\t" << real(entropy) << endl; + + return(real(entropy)); + } + */ + + /* + DP Entropy_rho (LiebLin_Bethe_State& RefState, int Delta) + { + // This function calculates the discrete entropy of a finite Lieb-Liniger state, + // counting the possible permutations within windows of fixed width. + + // We assume that the quantum numbers are ordered. + + // Calculate a density \rho_\Delta (x) for x_j on the lattice: + + int iK_UL = JSC::max(RefState.Ix2[0], RefState.Ix2[RefState.N - 1]); + + Vect rhoD(0.0, 2* iK_UL); + + int fourDeltasq = 4 * Delta * Delta; + for (int ix = 0; ix < 2* iK_UL; ++ix) { + // x = (-iK_UL + 1/2 + ix)/L + for (int i = 0; i < RefState.N; ++i) rhoD[ix] += 1.0/((-2*iK_UL + 1 + 2*ix -RefState.Ix2[i]) * (-2*iK_UL + 1 + 2*ix -RefState.Ix2[i]) + fourDeltasq); + rhoD[ix] *= 4.0 * Delta/PI; + //cout << "x = " << (-iK_UL + 1/2 + ix)/RefState.L << "\trhoD = " << rhoD[ix] << endl; + } + + // Now calculate the entropy: + DP entropy = 0.0; + for (int ix = 0; ix < 2* iK_UL; ++ix) + entropy -= rhoD[ix] * log(rhoD[ix]) + (1.0 - rhoD[ix]) * log(1.0 - rhoD[ix]); + + return(entropy); + } + */ + + DP Entropy_Fixed_Delta (LiebLin_Bethe_State& RefState, int Delta) + { + // This function calculates the discrete entropy of a finite Lieb-Liniger state, + // counting the possible permutations within windows of fixed width. + + // We assume that the quantum numbers are ordered. + + // Fill in vector of occupancies: + int nrIs = (RefState.Ix2[RefState.N-1] - RefState.Ix2[0])/2 + 1 + 2*Delta; // assume Ix2 are ordered, leave space of Delta on both sides + Vect occupancy(0, nrIs); // leave space of Delta on both sides + for (int i = 0; i < RefState.N; ++i) occupancy[Delta + (RefState.Ix2[i] -RefState.Ix2[0])/2] = 1; + // Check: + int ncheck = 0; + for (int i = 0; i < nrIs; ++i) if(occupancy[i] == 1) ncheck++; + //cout << "Check occupancy: " << endl << occupancy << endl; + if (ncheck != RefState.N) { + cout << ncheck << "\t" << RefState.N << endl; + JSCerror("Counting q numbers incorrectly in Entropy."); + } + + // Define some useful numbers: + Vect_DP oneoverDeltalnchoose(Delta + 1); + for (int i = 0; i <= Delta; ++i) oneoverDeltalnchoose[i] = ln_choose(Delta, i)/Delta; + //Vect_DP oneoverDeltalnchoosecheck(Delta + 1); + //for (int i = 0; i <= Delta; ++i) oneoverDeltalnchoosecheck[i] = log(DP(choose(Delta, i)))/Delta; + //cout << oneoverDeltalnchoose << endl; + //cout << oneoverDeltalnchoosecheck << endl; + + // Compute entropy: + DP entropy = 0.0; + int np; + for (int i = 0; i < nrIs - Delta; ++i) { + np = 0; + for (int iD = 0; iD < Delta; ++iD) if (occupancy[i + iD] == 1) np++; + + entropy += oneoverDeltalnchoose[np]; + } + + return(entropy); + } + + DP Entropy (LiebLin_Bethe_State& RefState, int Delta) + { + // Perform an average of entropies for regulators from Delta to 2Delta: + DP entropysum = 0.0; + for (int ie = 0; ie < Delta; ++ie) entropysum += Entropy_Fixed_Delta (RefState, Delta + ie); + return(entropysum/Delta); + } + + DP Entropy (LiebLin_Bethe_State& RefState) + { + int Delta = int(log(RefState.L)); + return(Entropy (RefState, Delta)); + } + /* + DP Canonical_Free_Energy (LiebLin_Bethe_State& RefState, DP kBT, int Delta) + { + return(RefState.E - kBT * Entropy (RefState, Delta)); + } + + DP Canonical_Free_Energy (LiebLin_Bethe_State& RefState, DP kBT, int Delta) + { + return(RefState.E - kBT * Entropy (RefState, Delta)); + } + */ + DP Canonical_Free_Energy (LiebLin_Bethe_State& RefState, DP kBT) + { + return(RefState.E - kBT * Entropy (RefState)); + } + + /* + //LiebLin_Bethe_State Canonical_Saddle_Point_State (DP c_int, DP L, int N, DP kBT, int Delta) + LiebLin_Bethe_State Canonical_Saddle_Point_State_pre20110618 (DP c_int, DP L, int N, DP kBT, DP epsilon) + { + // This function returns the discretized state minimizing the canonical free energy + // F = E - T S. + + LiebLin_Bethe_State spstate(c_int, L, N); + spstate.Compute_All (true); + + if (kBT < 1.0e-3) return(spstate); // interpret as zero T case + + LiebLin_Bethe_State spstateup = spstate; + LiebLin_Bethe_State spstatedown = spstate; + + bool converged = false; + bool convergedati = false; + + DP canfreeenstay, canfreeenup, canfreeendown; + DP Estay, Eup, Edown; + DP Sstay, Sup, Sdown; + + while (!converged) { + + spstate.Compute_All (false); + converged = true; // set to false if we change anything + + // If we can minimize the free energy by changing the quantum number, we do: + // NOTE: we keep the state symmetric w/r to parity. + + for (int i = 0; i < N/2; ++i) { + // Try to increase or decrease this quantum number + + convergedati = false; + + while (!convergedati) { + + convergedati = true; // set to false if we change anything + + Estay = spstate.E; + //Sstay = Entropy (spstate, Delta); + Sstay = Entropy (spstate, epsilon); + //canfreeenstay = Canonical_Free_Energy (spstate, kBT, Delta); + canfreeenstay = Estay - kBT * Sstay; + + spstateup = spstate; + if (i == 0 || spstateup.Ix2[i-1] < spstateup.Ix2[i] - 2) { + spstateup.Ix2[i] -= 2; + spstateup.Ix2[N-1-i] += 2; + spstateup.Compute_All(false); + Eup = spstateup.E; + //Sup = Entropy(spstateup, Delta); + Sup = Entropy(spstateup, epsilon); + //canfreeenup = Canonical_Free_Energy (spstateup, kBT, Delta); + canfreeenup = Eup - kBT * Sup; + } + else canfreeenup = canfreeenstay + 1.0e-6; + + spstatedown = spstate; + if (spstatedown.Ix2[i+1] > spstatedown.Ix2[i] + 2) { + spstatedown.Ix2[i] += 2; + spstatedown.Ix2[N-1-i] -= 2; + spstatedown.Compute_All(false); + Edown = spstatedown.E; + //Sdown = Entropy(spstatedown, Delta); + Sdown = Entropy(spstatedown, epsilon); + //canfreeendown = Canonical_Free_Energy (spstatedown, kBT, Delta); + canfreeendown = Edown - kBT * Sdown; + } + else canfreeendown = canfreeenstay + 1.0e-6; + + //cout << "i = " << i << "\t" << spstate.Ix2[i] << "\t\t" << canfreeenstay << "\t" << canfreeenup << "\t" << canfreeendown + // << "\t\t" << Estay << "\t" << Eup << "\t" << Edown << "\t\t" << Sstay << "\t" << Sup << "\t" << Sdown << endl; + // Choose what to do: + if (canfreeenup < canfreeenstay && canfreeendown < canfreeenstay) + cout << canfreeenstay << "\t" << canfreeenup << "\t" << canfreeendown << "\tWarning: unclear option for minimization." << endl; + else if (canfreeenup < canfreeenstay) { + spstate = spstateup; + convergedati = false; + converged = false; + } + else if (canfreeendown < canfreeenstay) { + spstate = spstatedown; + convergedati = false; + converged = false; + } + // else do nothing. + } // while !convergedati + + } // for i + } // while (!converged) + + return(spstate); + } + */ + + /* REMOVED FROM ++G_3 ONWARDS + //LiebLin_Bethe_State Canonical_Saddle_Point_State (DP c_int, DP L, int N, DP kBT, int Delta) + //LiebLin_Bethe_State Canonical_Saddle_Point_State (DP c_int, DP L, int N, DP kBT, DP epsilon) + LiebLin_Bethe_State Canonical_Saddle_Point_State_pre8 (DP c_int, DP L, int N, DP kBT) + { + // This function returns the discretized state minimizing the canonical free energy + // F = E - T S. + + // Improvement on version pre20110618: allow for pairwise `in' and `out' movements + // keeping the energy the same to order 1/L^2 but changing the entropy. + + // The regulator Delta also has become an `internal' parameter. + + LiebLin_Bethe_State spstate(c_int, L, N); + spstate.Compute_All (true); + + if (kBT < 1.0e-3) return(spstate); // interpret as zero T case + + LiebLin_Bethe_State spstateup = spstate; + LiebLin_Bethe_State spstatedown = spstate; + LiebLin_Bethe_State spstatein = spstate; + LiebLin_Bethe_State spstateout = spstate; + + bool converged = false; + bool convergedati = false; + + DP canfreeenstay, canfreeenup, canfreeendown, canfreeenin, canfreeenout; + DP Estay, Eup, Edown, Ein, Eout; + DP Sstay, Sup, Sdown, Sin, Sout; + Estay = 0.0; Eup = 0.0; Edown = 0.0; Ein = 0.0; Eout = 0.0; + Sstay = 0.0; Sup = 0.0; Sdown = 0.0; Sin = 0.0; Sout = 0.0; + + while (!converged) { + + spstate.Compute_All (false); + converged = true; // set to false if we change anything + + // If we can minimize the free energy by changing the quantum number, we do: + // NOTE: we keep the state symmetric w/r to parity. + + for (int i = 0; i < N/2 - 1; ++i) { + + // Try to increase or decrease the quantum numbers at i , + // or do an (approximately) energy-preserving move with i+1 (and parity pairs) + // giving 5 possible options: stay same, (\pm 1, 0), (+1, -1), (-1, +1) + + convergedati = false; + + while (!convergedati) { + + convergedati = true; // set to false if we change anything + + Estay = spstate.E; + Sstay = Entropy (spstate); + //Sstay = Entropy (spstate, Delta); + //Sstay = Entropy (spstate, epsilon); + //canfreeenstay = Canonical_Free_Energy (spstate, kBT, Delta); + canfreeenstay = Estay - kBT * Sstay; + + spstateup = spstate; + if (i == 0 || spstateup.Ix2[i-1] < spstateup.Ix2[i] - 2) { + spstateup.Ix2[i] -= 2; + spstateup.Ix2[N-1-i] += 2; + spstateup.Compute_All(false); + Eup = spstateup.E; + Sup = Entropy(spstateup); + //Sup = Entropy(spstateup, Delta); + //Sup = Entropy(spstateup, epsilon); + //canfreeenup = Canonical_Free_Energy (spstateup, kBT, Delta); + canfreeenup = Eup - kBT * Sup; + } + else canfreeenup = canfreeenstay + 1.0e-6; + + spstatedown = spstate; + if (spstatedown.Ix2[i+1] > spstatedown.Ix2[i] + 2) { + spstatedown.Ix2[i] += 2; + spstatedown.Ix2[N-1-i] -= 2; + spstatedown.Compute_All(false); + Edown = spstatedown.E; + Sdown = Entropy(spstatedown); + //Sdown = Entropy(spstatedown, Delta); + //Sdown = Entropy(spstatedown, epsilon); + //canfreeendown = Canonical_Free_Energy (spstatedown, kBT, Delta); + canfreeendown = Edown - kBT * Sdown; + } + else canfreeendown = canfreeenstay + 1.0e-6; + + spstatein = spstate; + if (spstatein.Ix2[i+1] > spstatein.Ix2[i] + 4) { // can move them closer + spstatein.Ix2[i] += 2; + spstatein.Ix2[i+1] -= 2; + spstatein.Ix2[N-1-i] -= 2; + spstatein.Ix2[N-1-i-1] += 2; + spstatein.Compute_All(false); + Ein = spstatein.E; + Sin = Entropy(spstatein); + //Sin = Entropy(spstatein, Delta); + //Sin = Entropy(spstatein, epsilon); + //canfreeenin = Canonical_Free_Energy (spstatein, kBT, Delta); + canfreeenin = Ein - kBT * Sin; + } + else canfreeenin = canfreeenstay + 1.0e-6; + + spstateout = spstate; + if (i == 0 && spstateout.Ix2[1] + 2 < spstateout.Ix2[2] + || (i < N/2 - 1 && spstateout.Ix2[i] - 2 > spstateout.Ix2[i-1] + && spstateout.Ix2[i+1] + 2 < spstateout.Ix2[i+2])) { // can move them further apart + spstateout.Ix2[i] -= 2; + spstateout.Ix2[i+1] += 2; + spstateout.Ix2[N-1-i] += 2; + spstateout.Ix2[N-1-i-1] -= 2; + spstateout.Compute_All(false); + Eout = spstateout.E; + Sout = Entropy(spstateout); + //Sout = Entropy(spstateout, Delta); + //Sout = Entropy(spstateout, epsilon); + //canfreeenout = Canonical_Free_Energy (spstateout, kBT, Delta); + canfreeenout = Eout - kBT * Sout; + } + else canfreeenout = canfreeenstay + 1.0e-6; + + //cout << setprecision(8) << "i = " << i << "\t" << spstate.Ix2[i] << "\t" << spstate.Ix2[i+1] << "\t\t" << canfreeenstay << "\t" << canfreeenup << "\t" << canfreeendown << "\t" << canfreeenin << "\t" << canfreeenout << endl; + //cout << "\t\tE: " << Estay << "\t" << Eup << "\t" << Edown << "\t" << Ein << "\t" << Eout << endl; + //cout << "\t\tS: " << Sstay << "\t" << Sup << "\t" << Sdown << "\t" << Sin << "\t" << Sout << endl; + + // Choose what to do: find minimum, + if (canfreeenstay < canfreeenup && canfreeenstay < canfreeendown && canfreeenstay < canfreeenin && canfreeenstay < canfreeenout) { + // do nothing, convergetati is already true + } + else if (canfreeenup < canfreeenstay && canfreeenup < canfreeendown && canfreeenup < canfreeenin && canfreeenup < canfreeenout) { + spstate = spstateup; + convergedati = false; + converged = false; + } + else if (canfreeendown < canfreeenstay && canfreeendown < canfreeenup && canfreeendown < canfreeenin && canfreeendown < canfreeenout) { + spstate = spstatedown; + convergedati = false; + converged = false; + } + else if (canfreeenin < canfreeenstay && canfreeenin < canfreeenup && canfreeenin < canfreeendown && canfreeenin < canfreeenout) { + spstate = spstatein; + convergedati = false; + converged = false; + } + else if (canfreeenout < canfreeenstay && canfreeenout < canfreeenup && canfreeenout < canfreeendown && canfreeenout < canfreeenin) { + spstate = spstateout; + convergedati = false; + converged = false; + } + else cout << canfreeenstay << "\t" << canfreeenup << "\t" << canfreeendown << "\t" << canfreeenin << "\t" << canfreeenout << "\tWarning: unclear option for minimization." << endl; + + // else do nothing. + } // while !convergedati + + } // for i + } // while (!converged) + + //cout << "Number of holes between I's: " << endl; + //for (int i = 0; i < N/2; ++i) cout << (spstate.Ix2[i+1] - spstate.Ix2[i])/2 - 1 << "\t"; + //cout << endl; + + return(spstate); + } + */ + + DP rho_of_lambdaoc_1 (LiebLin_Bethe_State& RefState, DP lambdaoc, DP delta) + { + DP answer = 0.0; + for (int i = 0; i < RefState.N; ++i) + answer += atan((lambdaoc - RefState.lambdaoc[i])/delta + 0.5) - atan((lambdaoc - RefState.lambdaoc[i])/delta - 0.5); + answer *= 1.0/(PI * delta * RefState.L); + + return(answer); + } + + DP rho_of_lambdaoc_2 (LiebLin_Bethe_State& RefState, DP lambdaoc, DP delta) + { + DP answer = 0.0; + for (int i = 0; i < RefState.N; ++i) + answer += 1.0/(pow(lambdaoc - RefState.lambdaoc[i], 2.0) + delta*delta); + answer *= delta/(PI * RefState.L); + + return(answer); + } + + // Better implementation: making use of rediscretized TBA state. + LiebLin_Bethe_State Canonical_Saddle_Point_State (DP c_int, DP L, int N, DP kBT) + { + // This function returns the discretized state minimizing the canonical free energy + // F = E - T S. + + // This is obtained by rediscretizing the solution coming from TBA. + + // ASSUMPTIONS: + // Periodic boundary conditions (the state which is output is forced to be symmetric Ix2 == -Ix2). + + // For zero temperature, return the ground state: + if (fabs(kBT) < 1.0e-4) return(LiebLin_Bethe_State(c_int, L, N)); + + // Otherwise, return the discretized TBA saddle-point state: + LiebLin_TBA_Solution TBAsol = LiebLin_TBA_Solution_fixed_nbar (c_int, N/L, kBT, 1.0e-4, JSC::max(N/10, 10)); + + LiebLin_Bethe_State spstate = Discretized_LiebLin_Bethe_State (c_int, L, N, TBAsol.rho); + + // Explicitly symmetrize: + for (int i = 0; i < N/2; ++i) spstate.Ix2[i] = -spstate.Ix2[N-1-i]; + spstate.Compute_All(false); + + return(spstate); + } + + LiebLin_Bethe_State Add_Particle_at_Center (const LiebLin_Bethe_State& RefState) + { + LiebLin_Bethe_State ReturnState (RefState.c_int, RefState.L, RefState.N + 1); + + // Add a quantum number in middle (explicitly: to right of index N/2) + // and shift quantum numbers by half-integer away from added one: + ReturnState.Ix2[RefState.N/2] = RefState.Ix2[RefState.N/2] - 1; + for (int i = 0; i < RefState.N+1; ++i) + ReturnState.Ix2[i + (i >= RefState.N/2)] = RefState.Ix2[i] - 1 + 2*(i >= RefState.N/2); + + return(ReturnState); + } + + LiebLin_Bethe_State Remove_Particle_at_Center (const LiebLin_Bethe_State& RefState) + { + LiebLin_Bethe_State ReturnState (RefState.c_int, RefState.L, RefState.N - 1); + + // Remove midmost and shift quantum numbers by half-integer towards removed one: + for (int i = 0; i < RefState.N-1; ++i) + ReturnState.Ix2[i] = RefState.Ix2[i + (i >= RefState.N/2)] + 1 - 2*(i >= RefState.N/2); + + return(ReturnState); + } + +} // namespace JSC diff --git a/src/LIEBLIN/LiebLin_Twisted_ln_Overlap.cc b/src/LIEBLIN/LiebLin_Twisted_ln_Overlap.cc new file mode 100644 index 0000000..fe8b693 --- /dev/null +++ b/src/LIEBLIN/LiebLin_Twisted_ln_Overlap.cc @@ -0,0 +1,117 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Twisted_ln_Overlap.cc + +Purpose: Calculates Nikita Slavnov's determinant, case RHS is Bethe + and LHS is twisted Bethe + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + complex Kernel_Twisted (complex expbeta, complex lambdaoc) + { + return(1.0/(lambdaoc + II) - expbeta/(lambdaoc - II)); + } + + complex Kernel_Twisted (complex expbeta, DP lambdaoc) + { + return(1.0/(lambdaoc + II) - expbeta/(lambdaoc - II)); + } + + complex Fn_V (int j, int sign, Vect >& lstate_lambdaoc, LiebLin_Bethe_State& rstate) + { + complex result_num = 1.0; + complex result_den = 1.0; + + complex signcx = complex(sign); + + for (int m = 0; m < rstate.N; ++m) { + result_num *= (lstate_lambdaoc[m] - rstate.lambdaoc[j] + signcx * II); + result_den *= (rstate.lambdaoc[m] - rstate.lambdaoc[j] + signcx * II); + } + + return(result_num/result_den); + } + + complex Fn_V_Nikita (int j, int sign, Vect >& lstate_lambdaoc, LiebLin_Bethe_State& rstate) + { + // To match with Nikita's new conventions + return(1.0/Fn_V (j, -sign, lstate_lambdaoc, rstate)); + } + + complex LiebLin_Twisted_ln_Overlap (DP expbeta, Vect lstate_lambdaoc, DP lstate_lnnorm, LiebLin_Bethe_State& rstate) + { + Vect > lstate_lambdaoc_CX(lstate_lambdaoc.size()); + for (int i = 0; i < lstate_lambdaoc.size(); ++i) lstate_lambdaoc_CX[i] = complex(lstate_lambdaoc[i]); + return(LiebLin_Twisted_ln_Overlap (complex(expbeta), lstate_lambdaoc_CX, lstate_lnnorm, rstate)); + } + + complex LiebLin_Twisted_ln_Overlap (complex expbeta, Vect > lstate_lambdaoc, DP lstate_lnnorm, LiebLin_Bethe_State& rstate) + { + // Computes the log of the overlap between the left state and the Bethe rstate + + // If momentum difference is zero but states are different, then form factor is zero: + + SQMat_CX one_plus_U (0.0, rstate.N); + + Vect_CX Vplus_Nikita (0.0, rstate.N); // contains V^+_j + Vect_CX Vminus_Nikita (0.0, rstate.N); // contains V^-_j + Vect_CX Fn_Prod (0.0, rstate.N); // product_{m\neq j} (\mu_m - \lambda_j)/(\lambda_m - \lambda_j) + Vect_CX rKern (0.0, rstate.N); // K(lambda_j - lambda_p) + + int p = 0; + + complex Kout = lstate_lambdaoc.sum() - rstate.K; + + for (int a = 0; a < rstate.N; ++a) { + Vplus_Nikita[a] = Fn_V_Nikita (a, 1, lstate_lambdaoc, rstate); + Vminus_Nikita[a] = Fn_V_Nikita (a, -1, lstate_lambdaoc, rstate); + Fn_Prod[a] = 1.0; + for (int m = 0; m < rstate.N; ++m) + if (m != a) Fn_Prod[a] *= (lstate_lambdaoc[m] - rstate.lambdaoc[a])/(rstate.lambdaoc[m] - rstate.lambdaoc[a]); + //rKern[a] = rstate.Kernel (a, p); + rKern[a] = Kernel_Twisted (expbeta, rstate.lambdaoc[p] - rstate.lambdaoc[a]); + } + + for (int a = 0; a < rstate.N; ++a) + for (int b = 0; b < rstate.N; ++b) + one_plus_U[a][b] = (a == b ? 1.0 : 0.0) + ((lstate_lambdaoc[a] - rstate.lambdaoc[a])/(1.0/Vplus_Nikita[a] - 1.0/Vminus_Nikita[a])) + * Fn_Prod[a] * (Kernel_Twisted(expbeta, rstate.lambdaoc[a] - rstate.lambdaoc[b]) - rKern[b]); + + complex ln_ddalpha_sigma = lndet_LU_CX_dstry(one_plus_U); + + complex ln_prod_V = 0.0; + for (int a = 0; a < rstate.N; ++a) ln_prod_V += log(expbeta * Vplus_Nikita[a]/Vminus_Nikita[a] - 1.0); + ln_prod_V -= 0.5 * rstate.N * log(expbeta); + + complex ln_prod_2 = 0.0; + for (int a = 0; a < rstate.N; ++a) + for (int b = 0; b < rstate.N; ++b) + ln_prod_2 += log((lstate_lambdaoc[a] - rstate.lambdaoc[b] - II)/(rstate.lambdaoc[a] - lstate_lambdaoc[b])); + + //cout << endl << ln_ddalpha_sigma << "\t" << ln_prod_V << "\t" << ln_prod_2 << "\t" << - log(2.0 * II * imag(Vplus[p])) << endl; + //cout << endl << ln_ddalpha_sigma << "\t" << ln_prod_V << "\t" << ln_prod_2 << "\t" << - log(Vplus[p] - Vminus[p]) << endl; + + ln_ddalpha_sigma += ln_prod_V + ln_prod_2 - log(Vplus_Nikita[p] - expbeta * Vminus_Nikita[p]); + + //cout << "shift = " << (complex(rstate.N) * (lstate_lambdaoc[0] - rstate.lambdaoc[0])/twoPI) << "\tKout = " << Kout << "\texp(-II*Kout) = " << exp(-II * Kout) + // << "\tlog(exp(-II * Kout) - 1.0) = " << log(exp(-II * Kout) - 1.0) << endl; + //cout << ln_ddalpha_sigma << "\t" << ln_prod_V << "\t" << ln_prod_2 << "\t" << lstate_lnnorm << endl << endl; + + //return (log(1.0 - expbeta) + ln_ddalpha_sigma - 0.5 * (lstate_lnnorm + rstate.lnnorm)); + return (log(1.0 - exp(-II * Kout)) + ln_ddalpha_sigma - 0.5 * (lstate_lnnorm + rstate.lnnorm)); + } + +} diff --git a/src/LIEBLIN/LiebLin_Twisted_lnnorm.cc b/src/LIEBLIN/LiebLin_Twisted_lnnorm.cc new file mode 100644 index 0000000..e5756e0 --- /dev/null +++ b/src/LIEBLIN/LiebLin_Twisted_lnnorm.cc @@ -0,0 +1,52 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Gaudin_lnnorm.cc + +Purpose: calculates the Gaudin norm of a vector of arbitrary + complex rapidities + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + DP LiebLin_Twisted_lnnorm (Vect >& lambdaoc, double cxL) + { + // Calculates the lnnorm of an eigenstate of the twisted transfer matrix + // so Gaudin can be used. + + int N = lambdaoc.size(); + + SQMat > Gaudin(N); + + complex sum_Kernel = 0.0; + + for (int j = 0; j < N; ++j) + for (int k = 0; k < N; ++k) { + if (j == k) { + sum_Kernel = 0.0; + for (int kp = 0; kp < N; ++kp) if (j != kp) sum_Kernel += 2.0/((lambdaoc[j] - lambdaoc[kp]) * (lambdaoc[j] - lambdaoc[kp]) + 1.0); + Gaudin[j][k] = cxL + sum_Kernel; + } + else Gaudin[j][k] = - 2.0/((lambdaoc[j] - lambdaoc[k]) * (lambdaoc[j] - lambdaoc[k]) + 1.0); + } + + DP lnnorm = real(lndet_LU_CX_dstry(Gaudin)); + + // Add the pieces outside of Gaudin determinant + for (int j = 0; j < N - 1; ++j) for (int k = j+1; k < N; ++k) lnnorm += log(1.0 + 1.0/norm(lambdaoc[j] - lambdaoc[k])); + + return(lnnorm); + } + +} // namespace JSC diff --git a/src/LIEBLIN/LiebLin_Utils.cc b/src/LIEBLIN/LiebLin_Utils.cc new file mode 100644 index 0000000..8ab64c0 --- /dev/null +++ b/src/LIEBLIN/LiebLin_Utils.cc @@ -0,0 +1,108 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: LiebLin_Utils.cc + +Purpose: Utilities for Lieb-Liniger gas + +***********************************************************/ + +#include "JSC.h" + +namespace JSC { + + DP LiebLin_dE0_dc (DP c_int, DP L, int N) + { + LiebLin_Bethe_State groundstate (c_int, L, N); + groundstate.Compute_All(true); + + DP dc_int = 1.0e-5 * c_int; + LiebLin_Bethe_State groundstate_2 (c_int + dc_int, L, N); + groundstate_2.Compute_All(true); + + return((groundstate_2.E - groundstate.E)/dc_int); + } + + DP LiebLin_vs (DP c_int, DP L, int N) + { + LiebLin_Bethe_State gstate (c_int, L, N); + gstate.Compute_All(true); + + DP Egs = gstate.E; + + gstate.Ix2[N-1] += 2; + + gstate.Compute_All(false); + + return((gstate.E - Egs) * L/twoPI); + + } + + DP LiebLin_Dressed_Charge_N (DP c_int, DP L, int N) + { + LiebLin_Bethe_State gstate(c_int, L, N); + gstate.Compute_All(true); + return(twoPI/(c_int * L * (gstate.lambdaoc[N-1] - gstate.lambdaoc[N-2]))); // lambda in units of c within code + } + + + // For computation of threshold exponents: + + int Momentum_Right_Excitations (LiebLin_Bethe_State& ScanState) + { + // Calculates the momentum of the excitations on the right of Fermi + // surface, discarding the rightmost excitation. + + int nr = 0; + for (int i = ScanState.N - 2; i >= ScanState.N/2; --i) + if (ScanState.Ix2[i] >= 0) nr += (ScanState.Ix2[i] + ScanState.N - 1 - 2*i)/2; + + return(nr); + } + + int Momentum_Left_Excitations (LiebLin_Bethe_State& ScanState) + { + // Calculates the momentum of the excitations on the left of Fermi + // surface, discarding the rightmost excitation. + + int nl = 0; + for (int i = 0; i < ScanState.N/2; ++i) + if (ScanState.Ix2[i] < 0) nl -= (ScanState.Ix2[i] + ScanState.N - 1 - 2*i)/2; + + return(nl); + } + + + DP ln_Overlap_with_BEC (LiebLin_Bethe_State& lambda) // Note: factorial approximated with Lanczos + { + int N = lambda.N; + int L = lambda.L; + + DP c_int = lambda.c_int; + + // The overlap identically vanishes if the state is not parity invariant: + if (!lambda.Check_Symmetry()) return(-300.0); + + SQMat_DP Gaudin( N); + + SQMat_DP Gaudin_Quench( N/2); + + lambda.Build_Reduced_Gaudin_Matrix ( Gaudin); + lambda.Build_Reduced_BEC_Quench_Gaudin_Matrix (Gaudin_Quench); + + //DP ln_prefactor = N/2.0 * log(2./fabs(c_int*L) ) + 0.5 * (N*log(N) - N + 0.5*log(2. * PI * N)); + + DP ln_prefactor = N/2.0 * log(2./(c_int*L) ) + 0.5 * real(ln_Gamma(complex(N+1.0))); + + for(int i =N/2; i Fn_V (int j, int sign, Vect >& lstate_lambdaoc, LiebLin_Bethe_State& rstate) + { + complex result_num = 1.0; + complex result_den = 1.0; + + complex signcx = complex(sign); + + for (int m = 0; m < rstate.N; ++m) { + result_num *= (lstate_lambdaoc[m] - rstate.lambdaoc[j] + signcx * II); + result_den *= (rstate.lambdaoc[m] - rstate.lambdaoc[j] + signcx * II); + } + + return(result_num/result_den); + } + */ + + complex LiebLin_ln_Overlap (Vect lstate_lambdaoc, DP lstate_lnnorm, LiebLin_Bethe_State& rstate) + { + Vect > lstate_lambdaoc_CX(rstate.N); + for (int i = 0; i < rstate.N; ++i) lstate_lambdaoc_CX[i] = complex(lstate_lambdaoc[i]); + return(LiebLin_ln_Overlap(lstate_lambdaoc_CX, lstate_lnnorm, rstate)); + } + + complex LiebLin_ln_Overlap (Vect > lstate_lambdaoc, DP lstate_lnnorm, LiebLin_Bethe_State& rstate) + { + // Computes the log of the overlap between the left state and the Bethe rstate + + // IMPORTANT ASSUMPTIONS: + // The length over which the overlap is calculated is that in rstate + + if (lstate_lambdaoc.size() != rstate.N) { + cout << "Caution: calling overlap of states with difference charges. Are you sure that's what you want ?" << endl; + return(0.0); + } + + // \mu are rstate rapidities, \lambda are from lstate (not BE) + + Vect_CX ln_prod_plus(0.0, rstate.N); // contains \prod_a (\mu_a - lambdaoc_k + II) + Vect_CX ln_prod_minus(0.0, rstate.N); // contains \prod_a (\mu_a - \lambdaoc_k - II) + Vect_CX ln_prod_ll_plus (0.0, rstate.N); // contains \prod_{a \neq k} (\lambdaoc_a - \lambdaoc_k + II) + Vect_CX ln_prod_ll_minus (0.0, rstate.N); // contains \prod_{a \neq k} (\lambdaoc_a - \lambdaoc_k - II) + + for (int j = 0; j < rstate.N; ++j) + for (int k = 0; k < rstate.N; ++k) { + ln_prod_plus[k] += log(rstate.lambdaoc[j] - lstate_lambdaoc[k] + II); + ln_prod_minus[k] += log(rstate.lambdaoc[j] - lstate_lambdaoc[k] - II); + if (j != k) { + ln_prod_ll_plus[k] += log(lstate_lambdaoc[j] - lstate_lambdaoc[k] + II); + ln_prod_ll_minus[k] += log(lstate_lambdaoc[j] - lstate_lambdaoc[k] - II); + } + } + + // Build the matrix: + SQMat_CX Omega (0.0, rstate.N); + + for (int j = 0; j < rstate.N; ++j) + for (int k = 0; k < rstate.N; ++k) + //Omega[j][k] = exp(-II * rstate.cxL * lstate_lambdaoc[k] + ln_prod_plus[k] - ln_prod_ll_plus[k]) + //* (II/((rstate.lambdaoc[j] - lstate_lambdaoc[k])*(rstate.lambdaoc[j] - lstate_lambdaoc[k] + II))) + //- (II/((rstate.lambdaoc[j] - lstate_lambdaoc[k])*(rstate.lambdaoc[j] - lstate_lambdaoc[k] - II))) + //* exp(ln_prod_minus[k] - ln_prod_ll_plus[k]); + Omega[j][k] = exp(-II * rstate.cxL * lstate_lambdaoc[k] - ln_prod_plus[k] + ln_prod_minus[k]) + * (II/((rstate.lambdaoc[j] - lstate_lambdaoc[k])*(rstate.lambdaoc[j] - lstate_lambdaoc[k] - II))) + - (II/((rstate.lambdaoc[j] - lstate_lambdaoc[k])*(rstate.lambdaoc[j] - lstate_lambdaoc[k] + II))); + + + //for (int j = 0; j < rstate.N; ++j) { + //for (int k = 0; k < rstate.N; ++k) + //cout << Omega[j][k] << "\t"; + //cout << endl; + //} + + + complex lndetOmega = lndet_LU_CX_dstry(Omega); + + //cout << "lndetOmega = " << lndetOmega << endl; + + // Prefactors: + complex ln_prod_d_mu = II * 0.5 * rstate.cxL * rstate.lambdaoc.sum(); + complex ln_prod_d_lambdaoc = II * 0.5 * rstate.cxL * lstate_lambdaoc.sum(); + + complex ln_prod_mu = 0.0; + complex ln_prod_lambdaoc = 0.0; + complex ln_prod_plusminus = 0.0; + for (int j = 0; j < rstate.N - 1; ++j) + for (int k = j+1; k < rstate.N; ++k) { + ln_prod_mu += log(rstate.lambdaoc[k] - rstate.lambdaoc[j]); + //ln_prod_lambdaoc += log((lstate_lambdaoc[j] - lstate_lambdaoc[k] + II) * (lstate_lambdaoc[j] - lstate_lambdaoc[k] - II)/(lstate_lambdaoc[k] - lstate_lambdaoc[j])); + ln_prod_lambdaoc += log(lstate_lambdaoc[k] - lstate_lambdaoc[j]); + } + + for (int j = 0; j < rstate.N; ++j) + for (int k = 0; k < rstate.N; ++k) + ln_prod_plusminus += log((rstate.lambdaoc[j] - lstate_lambdaoc[k] + II)); + + //cout << "ln_prod_mu " << ln_prod_mu << "\tln_prod_lambdaoc " << ln_prod_lambdaoc << "\tln_prod_plusminus " << ln_prod_plusminus + // << "\texp1 " << exp(-ln_prod_mu - ln_prod_lambdaoc) << "\texp2 " << exp(ln_prod_plusminus) << endl; + + //if (real(ln_prod_d_mu + ln_prod_d_lambdaoc - ln_prod_mu + ln_prod_lambdaoc + lndetOmega - 0.5 * (lstate_lnnorm + rstate.lnnorm)) > 10.0) { + //cout << ln_prod_d_mu << "\t" << ln_prod_d_lambdaoc << "\t" << -ln_prod_mu << "\t" << ln_prod_lambdaoc << "\t" << lndetOmega << "\t" << -0.5 * lstate_lnnorm << "\t" << -0.5 * rstate.lnnorm << endl; cout << ln_prod_d_mu + ln_prod_d_lambdaoc - ln_prod_mu + ln_prod_lambdaoc + lndetOmega - 0.5 * (lstate_lnnorm + rstate.lnnorm) << endl; + //JSCerror("Overlap exceeds 1."); + //} + + //return(ln_prod_d_mu + ln_prod_d_lambdaoc - ln_prod_mu + ln_prod_lambdaoc + lndetOmega - 0.5 * (lstate_lnnorm + rstate.lnnorm)); + return(ln_prod_d_mu + ln_prod_d_lambdaoc - ln_prod_mu - ln_prod_lambdaoc + ln_prod_plusminus + lndetOmega - 0.5 * (lstate_lnnorm + rstate.lnnorm)); + } + + /* + // Incorrect version + complex LiebLin_ln_Overlap (Vect > lstate_lambdaoc, DP lstate_lnnorm, LiebLin_Bethe_State& rstate) + { + // Computes the log of the overlap between the left state and the Bethe rstate + + // If momentum difference is zero but states are different, then form factor is zero: + + SQMat_CX one_plus_U (0.0, rstate.N); + + Vect_CX Vplus (0.0, rstate.N); // contains V^+_j + Vect_CX Vminus (0.0, rstate.N); // contains V^-_j + Vect_CX Fn_Prod (0.0, rstate.N); // product_{m\neq j} (\mu_m - \lambdaoc_j)/(\lambdaoc_m - \lambdaoc_j) + //Vect_DP rKern (0.0, rstate.N); // K(lambdaoc_j - lambdaoc_p) + + //int p = 0; + + complex Kout = lstate_lambdaoc.sum() - rstate.K; + + for (int a = 0; a < rstate.N; ++a) { + Vplus[a] = Fn_V (a, 1, lstate_lambdaoc, rstate); + Vminus[a] = Fn_V (a, -1, lstate_lambdaoc, rstate); + Fn_Prod[a] = 1.0; + for (int m = 0; m < rstate.N; ++m) + if (m != a) Fn_Prod[a] *= (lstate_lambdaoc[m] - rstate.lambdaoc[a])/(rstate.lambdaoc[m] - rstate.lambdaoc[a]); + //rKern[a] = rstate.Kernel (a, p); + } + + for (int a = 0; a < rstate.N; ++a) + for (int b = 0; b < rstate.N; ++b) + one_plus_U[a][b] = (a == b ? 1.0 : 0.0) + II * ((lstate_lambdaoc[a] - rstate.lambdaoc[a])/(Vplus[a] - Vminus[a])) + * Fn_Prod[a] * rstate.Kernel(a,b); + + complex ln_ddalpha_sigma = lndet_LU_CX_dstry(one_plus_U); + + complex ln_prod_V = 0.0; + for (int a = 0; a < rstate.N; ++a) ln_prod_V += log(Vplus[a] - Vminus[a]); + + complex ln_prod_2 = 0.0; + for (int a = 0; a < rstate.N; ++a) + for (int b = 0; b < rstate.N; ++b) + ln_prod_2 += log((rstate.lambdaoc[a] - rstate.lambdaoc[b] + II)/(lstate_lambdaoc[a] - rstate.lambdaoc[b])); + + //cout << endl << ln_ddalpha_sigma << "\t" << ln_prod_V << "\t" << ln_prod_2 << "\t" << - log(2.0 * II * imag(Vplus[p])) << endl; + cout << endl << ln_ddalpha_sigma << "\t" << ln_prod_V << "\t" << ln_prod_2 << "\t" << 0.5 * lstate_lnnorm << "\t" << 0.5 * rstate.lnnorm << endl;//- log(Vplus[p] - Vminus[p]) << endl; + + ln_ddalpha_sigma += ln_prod_V + ln_prod_2;// - log(Vplus[p] - Vminus[p]); + + //cout << "shift = " << (complex(rstate.N) * (lstate_lambdaoc[0] - rstate.lambdaoc[0])/twoPI) << "\tKout = " << Kout << "\texp(-II*Kout) = " << exp(-II * Kout) + // << "\tlog(exp(-II * Kout) - 1.0) = " << log(exp(-II * Kout) - 1.0) << endl; + //cout << ln_ddalpha_sigma << "\t" << ln_prod_V << "\t" << ln_prod_2 << "\t" << - log(Vplus[p] - Vminus[p]) << "\t" << lstate_lnnorm << endl << endl; + + return (ln_ddalpha_sigma - 0.5 * (lstate_lnnorm + rstate.lnnorm)); + } + */ + +} // namespace JSC diff --git a/src/LIEBLIN/ln_Density_ME.cc b/src/LIEBLIN/ln_Density_ME.cc new file mode 100644 index 0000000..3748c19 --- /dev/null +++ b/src/LIEBLIN/ln_Density_ME.cc @@ -0,0 +1,97 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: ln_Density_ME.cc + +Purpose: Computes the density operator \rho(x = 0) matrix element + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + complex Fn_V (int j, LiebLin_Bethe_State& lstate, LiebLin_Bethe_State& rstate) + { + //complex result_num = 1.0; + //complex result_den = 1.0; + complex result = 1.0; + + for (int m = 0; m < lstate.N; ++m) { + //result_num *= (lstate.lambdaoc[m] - rstate.lambdaoc[j] + II); + //result_den *= (rstate.lambdaoc[m] - rstate.lambdaoc[j] + II); + result *= (lstate.lambdaoc[m] - rstate.lambdaoc[j] + II)/(rstate.lambdaoc[m] - rstate.lambdaoc[j] + II); + } + + //return(result_num/result_den); + return(result); + } + + complex ln_Density_ME (LiebLin_Bethe_State& lstate, LiebLin_Bethe_State& rstate) + { + // Computes the log of the density operator \rho(x = 0) matrix element between lstate and rstate. + + // If we have lstate == rstate, density matrix element = N/L: + + if (lstate.Ix2 == rstate.Ix2) return(log(lstate.N/lstate.L)); + + // If momentum difference is zero but states are different, then matrix element is zero: + + else if (lstate.iK == rstate.iK) return(-200.0); // so exp(.) is zero + + SQMat_DP one_plus_U (0.0, lstate.N); + + Vect_CX Vplus (0.0, lstate.N); // contains V^+_j; V^-_j is the conjugate + //Vect_DP Fn_Prod (0.0, lstate.N); // product_{m\neq j} (\mu_m - \lambdaoc_j)/(\lambdaoc_m - \lambdaoc_j) + // From ABACUS++G_3 onwards: use logs to stabilize numerical values at small c, at the cost of execution speed. + Vect_CX ln_Fn_Prod (0.0, lstate.N); // product_{m\neq j} (\mu_m - \lambdaoc_j)/(\lambdaoc_m - \lambdaoc_j) + Vect_DP rKern (0.0, lstate.N); // K(lambdaoc_j - lambdaoc_(p == arbitrary)) + + //int p = 0; + int p = rstate.N/2-1; // choice doesn't matter, see 1990_Slavnov_TMP_82 after (3.8). Choose rapidity around the middle. + + DP Kout = lstate.K - rstate.K; + + for (int a = 0; a < lstate.N; ++a) { + Vplus[a] = Fn_V (a, lstate, rstate); + //Fn_Prod[a] = 1.0; + ln_Fn_Prod[a] = 0.0;; + for (int m = 0; m < lstate.N; ++m) + //if (m != a) Fn_Prod[a] *= (lstate.lambdaoc[m] - rstate.lambdaoc[a])/(rstate.lambdaoc[m] - rstate.lambdaoc[a]); + if (m != a) ln_Fn_Prod[a] += log(complex(lstate.lambdaoc[m] - rstate.lambdaoc[a])/(rstate.lambdaoc[m] - rstate.lambdaoc[a])); + rKern[a] = rstate.Kernel (a, p); + } + + for (int a = 0; a < lstate.N; ++a) + for (int b = 0; b < lstate.N; ++b) + one_plus_U[a][b] = (a == b ? 1.0 : 0.0) + 0.5 * ((lstate.lambdaoc[a] - rstate.lambdaoc[a])/imag(Vplus[a])) + //* Fn_Prod[a] * (rstate.Kernel(a,b) - rKern[b]); + * real(exp(ln_Fn_Prod[a])) * (rstate.Kernel(a,b) - rKern[b]); + + complex ln_ddalpha_sigma = lndet_LU_dstry(one_plus_U); + + complex ln_prod_V = 0.0; + for (int a = 0; a < lstate.N; ++a) ln_prod_V += log(2.0 * II * imag(Vplus[a])); + + complex ln_prod_2 = 0.0; + for (int a = 0; a < lstate.N; ++a) + for (int b = 0; b < lstate.N; ++b) + ln_prod_2 += log((rstate.lambdaoc[a] - rstate.lambdaoc[b] + II)/(lstate.lambdaoc[a] - rstate.lambdaoc[b])); + + //cout << "ln_Fn_Prod = " << ln_Fn_Prod << endl; + //cout << endl << ln_ddalpha_sigma << "\t" << ln_prod_V << "\t" << ln_prod_2 << "\t" << - log(2.0 * II * imag(Vplus[p])) << "\t" << - 0.5 * (lstate.lnnorm + rstate.lnnorm) << "\t" << ln_ddalpha_sigma + ln_prod_V + ln_prod_2 - log(2.0 * II * imag(Vplus[p])) - 0.5 * (lstate.lnnorm + rstate.lnnorm) << endl; + + ln_ddalpha_sigma += ln_prod_V + ln_prod_2 - log(2.0 * II * imag(Vplus[p])); + + return (log(-II * Kout) + ln_ddalpha_sigma - 0.5 * (lstate.lnnorm + rstate.lnnorm)); + } + +} diff --git a/src/LIEBLIN/ln_Psi_ME.cc b/src/LIEBLIN/ln_Psi_ME.cc new file mode 100644 index 0000000..851f748 --- /dev/null +++ b/src/LIEBLIN/ln_Psi_ME.cc @@ -0,0 +1,81 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: ln_Psi_ME.cc + +Purpose: Computes the matrix element of \Psi (x = 0) + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + complex Fn_V_Psi (int j, LiebLin_Bethe_State& lstate, LiebLin_Bethe_State& rstate) + { + // lstate has one particle less than rstate + + complex result_num = 1.0; + complex result_den = 1.0; + + for (int m = 0; m < lstate.N; ++m) + result_num *= (lstate.lambdaoc[m] - rstate.lambdaoc[j] + II); + + for (int m = 0; m < rstate.N; ++m) + result_den *= (rstate.lambdaoc[m] - rstate.lambdaoc[j] + II); + + return(result_num/result_den); + } + + complex ln_Psi_ME (LiebLin_Bethe_State& lstate, LiebLin_Bethe_State& rstate) + { + // Computes the log of the annihilation operator matrix element between lstate and rstate. + + if (lstate.N + 1 != rstate.N) JSCerror("Wrong particle numbers in left and right states for Psi FF."); + + SQMat_DP U_Psi (0.0, lstate.N); + + Vect_CX Vplus (0.0, lstate.N); // contains V^+_j; V^-_j is the conjugate + Vect_DP Fn_Prod (0.0, lstate.N); // product_{m} (\mu_m - \lambdaoc_j)/product_{m neq j} (\lambdaoc_m - \lambdaoc_j) + Vect_DP rKern (0.0, lstate.N); // K(lambdaoc_j - lambdaoc_(p == N)) + + int p = rstate.N - 1; + + for (int a = 0; a < lstate.N; ++a) { + Vplus[a] = Fn_V_Psi (a, lstate, rstate); + Fn_Prod[a] = (lstate.lambdaoc[a] - rstate.lambdaoc[a])/(rstate.lambdaoc[rstate.N - 1] - rstate.lambdaoc[a]); + for (int m = 0; m < lstate.N; ++m) + if (m != a) Fn_Prod[a] *= (lstate.lambdaoc[m] - rstate.lambdaoc[a])/(rstate.lambdaoc[m] - rstate.lambdaoc[a]); + rKern[a] = rstate.Kernel (a, p); + } + + for (int a = 0; a < lstate.N; ++a) + for (int b = 0; b < lstate.N; ++b) + U_Psi[a][b] = (a == b ? 2.0 * imag(Vplus[a]) : 0.0) + Fn_Prod[a] * (rstate.Kernel(a,b) - rKern[b]); + + complex ln_det_U_Psi = lndet_LU_dstry(U_Psi); + + complex ln_prod_lambdaocsq_plus_one = 0.0; + for (int a = 0; a < rstate.N - 1; ++a) + for (int b = a; b < rstate.N; ++b) + ln_prod_lambdaocsq_plus_one += log(complex((rstate.lambdaoc[a] - rstate.lambdaoc[b]) * (rstate.lambdaoc[a] - rstate.lambdaoc[b])) + 1.0); + + complex ln_prod_lambdaoca_min_mub = 0.0; + for (int a = 0; a < rstate.N; ++a) + for (int b = 0; b < lstate.N; ++b) + ln_prod_lambdaoca_min_mub += log(complex(rstate.lambdaoc[a] - lstate.lambdaoc[b])); + + //cout << endl << "Factors are " << ln_det_U_Psi << "\t" << ln_prod_lambdaocsq_plus_one << "\t" << ln_prod_lambdaoca_min_mub << endl; + + return (ln_det_U_Psi + 0.5 * log(lstate.c_int) + ln_prod_lambdaocsq_plus_one - ln_prod_lambdaoca_min_mub - 0.5 * (lstate.lnnorm + rstate.lnnorm)); + } + +} diff --git a/src/LIEBLIN/ln_g2_ME.cc b/src/LIEBLIN/ln_g2_ME.cc new file mode 100644 index 0000000..a65d1a7 --- /dev/null +++ b/src/LIEBLIN/ln_g2_ME.cc @@ -0,0 +1,193 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/LIEBLIN/ln_g2_ME.cc + +Purpose: provides the matrix element of the second density moment g2(x = 0) + +NOTE: based on Lorenzo Piroli's expression of the matrix element + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { +/* + complex ln_g2_ME_old_jacopo (LiebLin_Bethe_State& mu, LiebLin_Bethe_State& lambda) + { + + if (mu.Ix2 == lambda.Ix2) return(-200.); + + DP c_int = mu.c_int; + + SQMat_CX G(lambda.N); + SQMat_CX GpB(lambda.N); + SQMat_CX B(lambda.N); + + complex log_themp; + complex lnprefactor = 0. ; + + lnprefactor += 2.*log(c_int) ; + + for(int j=0; j < mu.N; ++j){ + for(int k=0; k < j; ++k){ + lnprefactor -= log((mu.lambdaoc[j] - mu.lambdaoc[k])); + lnprefactor -= log((lambda.lambdaoc[j] - lambda.lambdaoc[k])); + } + } + + + for(int j=0; j < mu.N; ++j){ + for(int k=0; k sum_n =0.; + for(int n=0; n < mu.N; ++n) { + // compute the matrices + for (int a = 0; a < mu.N; ++a){ + for (int b = 0; b < mu.N; ++b){ + if(b == n){ + G[a][b] = 2.*II* lambda.lambdaoc[a]; + } + else{ + G[a][b] = II*1./(lambda.lambdaoc[a] - mu.lambdaoc[b])*( VVP[b]*1.0/(mu.lambdaoc[b] - lambda.lambdaoc[a] + II) + + VVM[b]*1.0/(mu.lambdaoc[b] - lambda.lambdaoc[a] - II) ); + } + } + } + + for (int a = 0; a < mu.N; ++a){ + for (int b = 0; b < mu.N; ++b){ + if(b == n){ + B[a][b] = 0.; + } + else{ + B[a][b] = II*1.0/(mu.lambdaoc[b] - mu.lambdaoc[n] - II) ; + } + } + } + + for (int a = 0; a < mu.N; ++a){ + for (int b = 0; b < mu.N; ++b){ + GpB[a][b] = G[a][b] + B[a][b]; + } + } + //finally add + sum_n += exp( lndet_LU_CX(GpB) ) - exp(lndet_LU_CX(G) ) ; + } + + return(log(sum_n ) + lnprefactor - 0.5 * (mu.lnnorm + lambda.lnnorm)); + } +*/ + complex Fn_V_g2 (int j, LiebLin_Bethe_State& lstate, LiebLin_Bethe_State& rstate) + { + //complex result_num = 1.0; + //complex result_den = 1.0; + complex result = 1.0; + + for (int m = 0; m < lstate.N; ++m) { + //result_num *= (lstate.lambdaoc[m] - rstate.lambdaoc[j] + II); + //result_den *= (rstate.lambdaoc[m] - rstate.lambdaoc[j] + II); + result *= (lstate.lambdaoc[m] - rstate.lambdaoc[j] + II)/(rstate.lambdaoc[m] - rstate.lambdaoc[j] + II); + } + + //return(result_num/result_den); + return(result); + } + + + complex ln_g2_ME (LiebLin_Bethe_State& lstate, LiebLin_Bethe_State& rstate) + { + // Computes the log of the density operator \rho(x = 0) matrix element between lstate and rstate. + + // If we have lstate == rstate, density matrix element = 0: + + + if (lstate.Ix2 == rstate.Ix2) return(-200.); + + + + + SQMat_DP one_plus_U (0.0, lstate.N); + + Vect_CX Vplus (0.0, lstate.N); // contains V^+_j; V^-_j is the conjugate + //Vect_DP Fn_Prod (0.0, lstate.N); // product_{m\neq j} (\mu_m - \lambdaoc_j)/(\lambdaoc_m - \lambdaoc_j) + // From ABACUS++G_3 onwards: use logs to stabilize numerical values at small c, at the cost of execution speed. + Vect_CX ln_Fn_Prod (0.0, lstate.N); // product_{m\neq j} (\mu_m - \lambdaoc_j)/(\lambdaoc_m - \lambdaoc_j) + Vect_DP rKern (0.0, lstate.N); // K(lambdaoc_j - lambdaoc_(p == arbitrary)) + + //int p = 0; + int p = rstate.N/2-1; // choice doesn't matter, see 1990_Slavnov_TMP_82 after (3.8). Choose rapidity around the middle. + + + DP c_int = rstate.c_int; + + DP Eout = log(pow(lstate.E - rstate.E,2.)) - (2)*log(c_int) - log(2*lstate.N); + + for (int a = 0; a < lstate.N; ++a) { + Vplus[a] = Fn_V_g2 (a, lstate, rstate); + //Fn_Prod[a] = 1.0; + ln_Fn_Prod[a] = 0.0;; + for (int m = 0; m < lstate.N; ++m) + //if (m != a) Fn_Prod[a] *= (lstate.lambdaoc[m] - rstate.lambdaoc[a])/(rstate.lambdaoc[m] - rstate.lambdaoc[a]); + if (m != a) ln_Fn_Prod[a] += log(complex(lstate.lambdaoc[m] - rstate.lambdaoc[a])/(rstate.lambdaoc[m] - rstate.lambdaoc[a])); + rKern[a] = rstate.Kernel (a, p); + } + + for (int a = 0; a < lstate.N; ++a) + for (int b = 0; b < lstate.N; ++b) + one_plus_U[a][b] = (a == b ? 1.0 : 0.0) + 0.5 * 1./imag(Vplus[a]) + //* Fn_Prod[a] * (rstate.Kernel(a,b) - rKern[b]); + * ((lstate.lambdaoc[a] - rstate.lambdaoc[a])*real(exp(ln_Fn_Prod[a])) * (rstate.Kernel(a,b) - rKern[b]) + rKern[b]); + + complex ln_ddalpha_sigma = lndet_LU_dstry(one_plus_U); + + complex ln_prod_V = 0.0; + for (int a = 0; a < lstate.N; ++a) ln_prod_V += log(2.0 * II * imag(Vplus[a])); + + complex ln_prod_2 = 0.0; + for (int a = 0; a < lstate.N; ++a) + for (int b = 0; b < lstate.N; ++b) + ln_prod_2 += log((rstate.lambdaoc[a] - rstate.lambdaoc[b] + II)/(lstate.lambdaoc[a] - rstate.lambdaoc[b])); + + //cout << "ln_Fn_Prod = " << ln_Fn_Prod << endl; + //cout << endl << ln_ddalpha_sigma << "\t" << ln_prod_V << "\t" << ln_prod_2 << "\t" << - log(2.0 * II * imag(Vplus[p])) << "\t" << - 0.5 * (lstate.lnnorm + rstate.lnnorm) << "\t" << ln_ddalpha_sigma + ln_prod_V + ln_prod_2 - log(2.0 * II * imag(Vplus[p])) - 0.5 * (lstate.lnnorm + rstate.lnnorm) << endl; + + ln_ddalpha_sigma += ln_prod_V + ln_prod_2 - log(2.0 * II * imag(Vplus[p])); + + return (log(II) + Eout + ln_ddalpha_sigma - 0.5 * (lstate.lnnorm + rstate.lnnorm)); + } + + + + + +} // namespace JSC diff --git a/src/MATRIX/balanc.cc b/src/MATRIX/balanc.cc new file mode 100644 index 0000000..a967b80 --- /dev/null +++ b/src/MATRIX/balanc.cc @@ -0,0 +1,47 @@ +#include "JSC.h" +using namespace std; + +void JSC::balanc(SQMat_DP& a) +{ + // Balancing of nonsymmetric matrix. See NR p.489. + + const DP RADIX = numeric_limits::radix; + + int i, j, last = 0; + DP s, r, g, f, c, sqrdx; + + int n = a.size(); + sqrdx = RADIX * RADIX; + + while (last == 0) { + last = 1; + for (i = 0; i < n; i++) { + r = c = 0.0; + for (j = 0; j < n; j++) + if (j != i) { + c += fabs(a[j][i]); + r += fabs(a[i][j]); + } + if (c != 0.0 && r != 0.0) { + g = r/RADIX; + f = 1.0; + s = c + r; + while (c < g) { + f *= RADIX; + c *= sqrdx; + } + g = r * RADIX; + while (c > g) { + f /= RADIX; + c /= sqrdx; + } + if ((c + r)/f < 0.95 * s) { + last = 0; + g = 1.0/f; + for (j = 0; j < n; j++) a[i][j] *= g; + for (j = 0; j < n; j++) a[j][i] *= f; + } + } + } + } +} diff --git a/src/MATRIX/det_LU.cc b/src/MATRIX/det_LU.cc new file mode 100644 index 0000000..8fb600a --- /dev/null +++ b/src/MATRIX/det_LU.cc @@ -0,0 +1,20 @@ +#include "JSC.h" +using namespace std; + +DP JSC::det_LU (SQMat_DP a) +{ + // Returns the determinant of matrix a, through LU decomposition + // In order to preserve the original matrix, it is copied first. + + Vect_INT indx(a.size()); + + SQMat_DP mat = a; + + DP d; + + JSC::ludcmp (mat, indx, d); + + for (int j = 0; j < mat.size(); j++) d *= mat[j][j]; + + return(d); +} diff --git a/src/MATRIX/det_LU_CX.cc b/src/MATRIX/det_LU_CX.cc new file mode 100644 index 0000000..3aa0510 --- /dev/null +++ b/src/MATRIX/det_LU_CX.cc @@ -0,0 +1,21 @@ +#include "JSC.h" +using namespace std; + +complex JSC::det_LU_CX (SQMat_CX a) +{ + // Returns the determinant of matrix a, through LU decomposition + + Vect_INT indx(a.size()); + + SQMat_CX mat = a; + + DP d; + + JSC::ludcmp_CX (mat, indx, d); + + complex dd = d; + + for (int j = 0; j < mat.size(); j++) dd *= mat[j][j]; + + return(dd); +} diff --git a/src/MATRIX/eigsrt.cc b/src/MATRIX/eigsrt.cc new file mode 100644 index 0000000..1ee9c93 --- /dev/null +++ b/src/MATRIX/eigsrt.cc @@ -0,0 +1,28 @@ +#include "JSC.h" + +void JSC::eigsrt (Vect_DP& d, SQMat_DP& v) +{ + // This puts the eigenvalues in INCREASING order, not decreasing as in NR ! + + int i, j, k; + DP p; + + int n = d.size(); + for (i = 0; i < n-1; i++) { + p = d[k=i]; + for (j = i; j < n; j++) { + if (d[j] <= p) { + p = d[k = j]; + } + } + if (k != i) { + d[k] = d[i]; + d[i] = p; + for (j = 0; j < n; j++) { + p = v[j][i]; + v[j][i] = v[j][k]; + v[j][k] = p; + } + } + } +} diff --git a/src/MATRIX/elmhes.cc b/src/MATRIX/elmhes.cc new file mode 100644 index 0000000..6b72ac2 --- /dev/null +++ b/src/MATRIX/elmhes.cc @@ -0,0 +1,36 @@ +#include "JSC.h" +using namespace std; + +void JSC::elmhes(SQMat_DP& a) +{ + // Reduction to Hessenberg form by elimination method. NR p.490 + + int i, j, m; + DP y, x; + + int n = a.size(); + for (m = 1; m < n-1; m++) { + x = 0.0; + i = m; + for (j = m; j < n; j++) { + if (fabs(a[j][m-1]) > fabs(x)) { + x = a[j][m-1]; + i = j; + } + } + if (i != m) { + for (j = m - 1; j < n; j++) SWAP(a[i][j], a[m][j]); + for (j = 0; j < n; j++) SWAP(a[j][i], a[j][m]); + } + if (x != 0.0) { + for (i = m + 1; i < n; i++) { + if ((y = a[i][m-1]) != 0.0) { + y /= x; + a[i][m-1] = y; + for (j = m; j < n; j++) a[i][j] -= y*a[m][j]; + for (j = 0; j < n; j++) a[j][m] += y*a[j][i]; + } + } + } + } +} diff --git a/src/MATRIX/gaussj.cc b/src/MATRIX/gaussj.cc new file mode 100644 index 0000000..8e1a53e --- /dev/null +++ b/src/MATRIX/gaussj.cc @@ -0,0 +1,56 @@ +#include "JSC.h" +using namespace std; + +void JSC::gaussj (SQMat_DP& a, SQMat_DP& b) +{ + int i, j, k, l, ll; + int icol = 0; + int irow = 0; + DP big, dum, pivinv; + + int n = a.size(); + int m = b.size(); + Vect_INT indxc(n), indxr(n), ipiv(n); + for (j = 0; j < n; j++) ipiv[j] = 0; + for (i = 0; i < n; i++) { + big = 0.0; + for (j = 0; j < n; j++) + if (ipiv[j] != 1) + for (k = 0; k < n; k++) { + if (ipiv[k] == 0) { + if (fabs(a[j][k]) >= big) { + big = fabs(a[j][k]); + irow = j; + icol = k; + } + } + } + ++(ipiv[icol]); + + if (irow != icol) { + for (l = 0; l < n; l++) SWAP (a[irow][l], a[icol][l]); + for (l = 0; l < m; l++) SWAP (b[irow][l], b[icol][l]); + } + indxr[i] = irow; + indxc[i] = icol; + if (a[icol][icol] == 0.0) JSCerror("gaussj: singular matrix"); + pivinv = 1.0/a[icol][icol]; + a[icol][icol] = 1.0; + for (l = 0; l < n; l++) a[icol][l] *= pivinv; + for (l = 0; l < m; l++) b[icol][l] *= pivinv; + for (ll = 0; ll < n; ll++) + if (ll != icol) { + dum = a[ll][icol]; + a[ll][icol] = 0.0; + for (l = 0; l < n; l++) a[ll][l] -= a[icol][l] * dum; + for (l = 0; l < m; l++) b[ll][l] -= b[icol][l] * dum; + } + } + + for (l = n - 1; l >= 0; l--) { + if (indxr[l] != indxc[l]) + for (k = 0; k < n; k++) + SWAP (a[k][indxr[l]], a[k][indxc[l]]); + } +} + diff --git a/src/MATRIX/hqr.cc b/src/MATRIX/hqr.cc new file mode 100644 index 0000000..5186d8a --- /dev/null +++ b/src/MATRIX/hqr.cc @@ -0,0 +1,131 @@ +#include "JSC.h" +using namespace std; + +void JSC::hqr(SQMat_DP& a, Vect_CX& wri) +{ + // Finds all eigenvalues of an upper Hessenberg matrix + + int nn, m, l, k, j, its, i, mmin; + DP z, y, x, w, v, u, t, s, anorm; + DP r = 0.0; + DP p = 0.0; + DP q = 0.0; + + int n = a.size(); + anorm = 0.0; + for (i = 0; i < n; ++i) + for (j = JSC::max(i-1, 0); j < n; j++) anorm += fabs(a[i][j]); + nn = n-1; + t = 0.0; + while (nn >= 0) { + its = 0; + do { + for (l = nn; l > 0; l--) { + s = fabs(a[l-1][l-1]) + fabs(a[l][l]); + if (s == 0.0) s = anorm; + if (fabs(a[l][l-1]) + s == s) { + a[l][l-1] = 0.0; + break; + } + } + x = a[nn][nn]; + if (l == nn) { + wri[nn--] = x + t; + } else { + y = a[nn-1][nn-1]; + w = a[nn][nn-1] * a[nn-1][nn]; + if (l == nn-1) { + p = 0.5*(y-x); + q = p*p + w; + z = sqrt(fabs(q)); + x += t; + if (q >= 0.0) { + z = p + SIGN(z, p); + wri[nn-1] = wri[nn] = x + z; + if (z != 0.0) wri[nn] = x - w/z; + } else { + wri[nn] = complex(x + p, z); + wri[nn-1] = conj(wri[nn]); + } + nn -=2; + } else { + if (its == 30) JSCerror("Too many iterations in hqr."); + if (its == 10 || its == 20) { + t += x; + for (i = 0; i < nn + 1; i++) a[i][i] -= x; + s = fabs(a[nn][nn-1]) + fabs(a[nn-1][nn-2]); + y = x = 0.75 * s; + w = -0.4375 * s * s; + } + ++its; + for (m = nn-2; m >= l; m--) { + z = a[m][m]; + r = x - z; + s = y - z; + p = (r*s - w)/a[m+1][m] + a[m][m+1]; + q = a[m+1][m+1] - z - r - s; + r = a[m+2][m+1]; + s = fabs(p) + fabs(q) + fabs(r); + p /= s; + q /= s; + r /= s; + if (m == l) break; + u = fabs(a[m][m-1]) * (fabs(q) + fabs(r)); + v = fabs(p) * (fabs(a[m-1][m-1]) + fabs(z) + fabs(a[m+1][m+1])); + if (u + v == v) break; + } + for (i = m; i < nn-1; i++) { + a[i+2][i] = 0.0; + if (i != m) a[i+2][i-1] = 0.0; + } + for (k = m; k < nn; k++) { + if (k != m) { + p = a[k][k-1]; + q = a[k+1][k-1]; + r = 0.0; + if (k+1 != nn) r = a[k+2][k-1]; + if ((x = fabs(p) + fabs(q) + fabs(r)) != 0.0) { + p /= x; + q /= x; + r /= x; + } + } + if ((s = SIGN(sqrt(p*p + q*q + r*r), p)) != 0.0) { + if (k == m) { + if (l != m) a[k][k-1] = -a[k][k-1]; + } + else a[k][k-1] = -s*x; + p += s; + x = p/s; + y = q/s; + z = r/s; + q /= p; + r /= p; + for (j = k; j < nn + 1; j++) { + p = a[k][j] + q* a[k+1][j]; + if (k+1 != nn) { + p += r*a[k+2][j]; + a[k+2][j] -= p*z; + } + a[k+1][j] -= p*y; + a[k][j] -= p*x; + } + mmin = nn < k+3 ? nn : k+3; + for (i = l; i < mmin+1; i++) { + p = x*a[i][k] + y*a[i][k+1]; + if (k != (nn)) { + p += z*a[i][k+2]; + a[i][k+2] -= p*r; + } + a[i][k+1] -= p*q; + a[i][k] -= p; + } + } + } + } + } + } while (l+1 < nn); + } +} + + diff --git a/src/MATRIX/jacobi.cc b/src/MATRIX/jacobi.cc new file mode 100644 index 0000000..ca0690e --- /dev/null +++ b/src/MATRIX/jacobi.cc @@ -0,0 +1,94 @@ +#include "JSC.h" +using namespace std; +using namespace JSC; + +namespace { + + inline void rot (SQMat& a, const DP s, const DP tau, const int i, const int j, + const int k, const int l) + { + DP g, h; + + g = a[i][j]; + h = a[k][l]; + a[i][j] = g - s * (h + g * tau); + a[k][l] = h + s * (g - h * tau); + } + +} + +void JSC::jacobi (SQMat& a, Vect_DP& d, SQMat& v, int& nrot) +{ + int i, j, ip, iq; + DP tresh, theta, tau, t, sm, s, h, g, c; + + int n = a.size(); // + Vect_DP b(n), z(n); + for (ip = 0; ip < n; ++ip) { // initialize to identity + for (iq = 0; iq < n; ++iq) v[ip][iq] = 0.0; + v[ip][ip] = 1.0; + } + + for (ip = 0; ip < n; ++ip) { + b[ip] = d[ip] = a[ip][ip]; + z[ip] = 0.0; + } + + nrot = 0; + + for (i = 1; i <= 50; i++) { + cout << "Sweep number " << i << " under way..." << endl; + sm = 0.0; + for (ip = 0; ip < n-1; ip++) { // sum off-diagonal elements + for (iq = ip + 1; iq < n; iq++) + sm += fabs(a[ip][iq]); + } + + if (sm == 0.0) return; + + if (i < 4) tresh = 0.2 * sm/(n*n); + else tresh = 0.0; + + for (ip = 0; ip < n-1; ip++) { + for (iq = ip + 1; iq < n; iq++) { + g = 100.0 * fabs(a[ip][iq]); + if ((i > 4) && (fabs(d[ip]) + g) == fabs(d[ip]) && (fabs(d[iq]) + g) == fabs(d[iq])) + a[ip][iq] = 0.0; + else if (fabs(a[ip][iq]) > tresh) { + h = d[iq] - d[ip]; + if (( fabs(h) + g) == fabs(h)) + t = (a[ip][iq])/h; + else { + theta = 0.5*h/(a[ip][iq]); + t = 1.0/(fabs(theta) + sqrt(1.0 + theta*theta)); + if (theta < 0.0) t = -t; + } + c = 1.0/sqrt(1.0 + t*t); + s = t*c; + tau = s/(1.0 + c); + h = t*a[ip][iq]; + z[ip] -= h; + z[iq] += h; + d[ip] -= h; + d[iq] += h; + a[ip][iq] = 0.0; + + for (j = 0; j < ip; j++) rot(a, s, tau, j, ip, j, iq); + for (j = ip+1; j < iq; j++) rot(a, s, tau, ip, j, j, iq); + for (j = iq+1; j < n; j++) rot(a, s, tau, ip, j, iq, j); + for (j = 0; j < n; j++) rot(v, s, tau, j, ip, j, iq); + + ++nrot; + } + } + } + + for (ip = 0; ip < n; ip++) { + b[ip] += z[ip]; + d[ip] = b[ip]; + z[ip] = 0.0; + } + } + cout << "Too many iterations in routine jacobi." << endl; +} + diff --git a/src/MATRIX/lndet_LU.cc b/src/MATRIX/lndet_LU.cc new file mode 100644 index 0000000..c4dcd55 --- /dev/null +++ b/src/MATRIX/lndet_LU.cc @@ -0,0 +1,22 @@ +#include "JSC.h" +using namespace std; + +DP JSC::lndet_LU (SQMat_DP a) +{ + // Returns the ln of abs of determinant of matrix a, through LU decomposition + + Vect_INT indx(a.size()); + + SQMat_DP mat = a; + + DP d; + + JSC::ludcmp (mat, indx, d); + + DP logdet = log(DP(d)); + for (int j = 0; j < mat.size(); j++) { + logdet += log(fabs(mat[j][j])); + } + + return(logdet); +} diff --git a/src/MATRIX/lndet_LU_CX.cc b/src/MATRIX/lndet_LU_CX.cc new file mode 100644 index 0000000..ecc40f1 --- /dev/null +++ b/src/MATRIX/lndet_LU_CX.cc @@ -0,0 +1,32 @@ +#include "JSC.h" +using namespace std; + +complex JSC::lndet_LU_CX (SQMat_CX a) +{ + // Returns the ln of determinant of matrix a, through LU decomposition + + Vect_INT indx(a.size()); + + SQMat_CX mat = a; + + DP d; + + complex logdet = 0.0; + + try { + + JSC::ludcmp_CX (mat, indx, d); + + logdet = log(complex(d)); + + for (int j = 0; j < mat.size(); j++) { + logdet += log(mat[j][j]); + } + } + + catch (Divide_by_zero) { + logdet = log(-1.0); // reset to nan + } + + return(logdet); +} diff --git a/src/MATRIX/lndet_LU_CX_dstry.cc b/src/MATRIX/lndet_LU_CX_dstry.cc new file mode 100644 index 0000000..4ec2ac2 --- /dev/null +++ b/src/MATRIX/lndet_LU_CX_dstry.cc @@ -0,0 +1,31 @@ +#include "JSC.h" +using namespace std; + +complex JSC::lndet_LU_CX_dstry (SQMat_CX& a) +{ + // Returns the ln of determinant of matrix a, through LU decomposition + // Does not preserve a, unlike lndet_LU_CX + + Vect_INT indx(a.size()); + + DP d; + + complex logdet = 0.0; + + try { + + JSC::ludcmp_CX (a, indx, d); + + logdet = log(complex(d)); + + for (int j = 0; j < a.size(); j++) { + logdet += log(a[j][j]); + } + } + + catch (Divide_by_zero) { + logdet = log(-1.0); // reset to nan + } + + return(logdet); +} diff --git a/src/MATRIX/lndet_LU_dstry.cc b/src/MATRIX/lndet_LU_dstry.cc new file mode 100644 index 0000000..086b304 --- /dev/null +++ b/src/MATRIX/lndet_LU_dstry.cc @@ -0,0 +1,31 @@ +#include "JSC.h" +using namespace std; + +complex JSC::lndet_LU_dstry (SQMat_DP& a) +{ + // Returns the ln of determinant of matrix a, through LU decomposition + // Does not preserve a, unlike lndet_LU + + Vect_INT indx(a.size()); + + DP d; + + complex logdet = 0.0; + + try { + + JSC::ludcmp (a, indx, d); + + logdet = log(complex(d)); + + for (int j = 0; j < a.size(); j++) { + logdet += log(complex(a[j][j])); + } + } + + catch (Divide_by_zero) { + logdet = log(-1.0); // reset to nan + } + + return(logdet); +} diff --git a/src/MATRIX/lubksb.cc b/src/MATRIX/lubksb.cc new file mode 100644 index 0000000..5fb33d0 --- /dev/null +++ b/src/MATRIX/lubksb.cc @@ -0,0 +1,25 @@ +#include "JSC.h" +using namespace std; + +void JSC::lubksb (SQMat_DP& a, Vect_INT& indx, Vect_DP& b) +{ + int i, ii=0, ip, j; + DP sum; + + int n = a.size(); + for (i = 0; i < n; i++) { + ip = indx[i]; + sum = b[ip]; + b[ip] = b[i]; + if (ii != 0) + for (j = ii-1; j < i; j++) sum -= a[i][j] * b[j]; + else if (sum != 0.0) + ii = i + 1; + b[i] = sum; + } + for (i = n - 1; i >= 0; i--) { + sum = b[i]; + for (j = i + 1; j < n; j++) sum -= a[i][j] * b[j]; + b[i] = sum/a[i][i]; + } +} diff --git a/src/MATRIX/lubksb_CX.cc b/src/MATRIX/lubksb_CX.cc new file mode 100644 index 0000000..83be6c7 --- /dev/null +++ b/src/MATRIX/lubksb_CX.cc @@ -0,0 +1,25 @@ +#include "JSC.h" +using namespace std; + +void JSC::lubksb_CX (SQMat_CX& a, Vect_INT& indx, Vect_CX& b) +{ + int i, ii=0, ip, j; + complex sum; + + int n = a.size(); + for (i = 0; i < n; i++) { + ip = indx[i]; + sum = b[ip]; + b[ip] = b[i]; + if (ii != 0) + for (j = ii-1; j < i; j++) sum -= a[i][j] * b[j]; + else if (sum != 0.0) + ii = i + 1; + b[i] = sum; + } + for (i = n - 1; i >= 0; i--) { + sum = b[i]; + for (j = i + 1; j < n; j++) sum -= a[i][j] * b[j]; + b[i] = sum/a[i][i]; + } +} diff --git a/src/MATRIX/ludcmp.cc b/src/MATRIX/ludcmp.cc new file mode 100644 index 0000000..5eec43c --- /dev/null +++ b/src/MATRIX/ludcmp.cc @@ -0,0 +1,55 @@ +#include "JSC.h" +using namespace std; + +void JSC::ludcmp (SQMat_DP& a, Vect_INT& indx, DP& d) +{ + const DP TINY = 1.0e-200; + int i, j, k; + int imax = 0; + DP big, dum, sum, temp; + + int n = a.size(); + if (n == 0) JSCerror("Zero-sized array in ludcmp"); + Vect_DP vv(n); + d = 1.0; + for (i = 0; i < n; i++) { + big = 0.0; + for (j = 0; j < n; j++) + if ((temp = fabs(a[i][j])) > big) big = temp; + if (big == 0.0) throw Divide_by_zero(); //JSCerror("Singular matrix in routine ludcmp."); + vv[i] = 1.0/big; + } + for (j = 0; j < n; j++) { + for (i = 0; i < j; i++) { + sum = a[i][j]; + for (k = 0; k < i; k++) sum -= a[i][k] * a[k][j]; + a[i][j] = sum; + } + big = 0.0; + for (i = j; i < n; i++) { + sum = a[i][j]; + for (k = 0; k < j; k++) sum -= a[i][k] * a[k][j]; + a[i][j] = sum; + if ((dum = vv[i]*fabs(sum)) >= big) { + big = dum; + imax = i; + } + } + if (j != imax) { + for (k = 0; k < n; k++) { + dum = a[imax][k]; + a[imax][k] = a[j][k]; + a[j][k] = dum; + } + d = -d; + vv[imax] = vv[j]; + } + indx[j] = imax; + if (a[j][j] == 0.0) a[j][j] = TINY; + if (j != n-1) { + dum = 1.0/(a[j][j]); + for (i = j + 1; i < n; i++) a[i][j] *= dum; + } + } +} + diff --git a/src/MATRIX/ludcmp_CX.cc b/src/MATRIX/ludcmp_CX.cc new file mode 100644 index 0000000..9fbe2f4 --- /dev/null +++ b/src/MATRIX/ludcmp_CX.cc @@ -0,0 +1,60 @@ +#include "JSC.h" +using namespace std; + +void JSC::ludcmp_CX (SQMat_CX& a, Vect_INT& indx, DP& d) +{ + const complex TINY = 1.0e-200; + int i, j, k; + int imax = 0; + // DP big, dum, sum, temp; + complex big, dum, sum, temp; + + int n = a.size(); + if (n == 0) JSCerror("Zero-sized array in ludcmp_CX"); + Vect_CX vv(n); + d = 1.0; + for (i = 0; i < n; i++) { + big = 0.0; + for (j = 0; j < n; j++) + // if ((temp = fabs(a[i][j])) > big) big = temp; + if ((abs(temp = a[i][j])) > abs(big)) big = temp; + //if ((norm(temp = a[i][j])) > norm(big)) big = temp; + if (big == 0.0) throw Divide_by_zero(); //JSCerror("Singular matrix in routine ludcmp."); + vv[i] = 1.0/big; + } + for (j = 0; j < n; j++) { + for (i = 0; i < j; i++) { + sum = a[i][j]; + for (k = 0; k < i; k++) sum -= a[i][k] * a[k][j]; + a[i][j] = sum; + } + big = 0.0; + for (i = j; i < n; i++) { + sum = a[i][j]; + for (k = 0; k < j; k++) sum -= a[i][k] * a[k][j]; + a[i][j] = sum; + // if ((dum = vv[i]*fabs(sum)) >= big) { + if ((abs(dum = vv[i]*sum)) >= abs(big)) { + //if ((norm(dum = vv[i]*sum)) >= norm(big)) { + big = dum; + imax = i; + } + } + if (j != imax) { + for (k = 0; k < n; k++) { + dum = a[imax][k]; + a[imax][k] = a[j][k]; + a[j][k] = dum; + } + d = -d; + vv[imax] = vv[j]; + } + indx[j] = imax; + if (a[j][j] == 0.0) a[j][j] = TINY; + if (j !=n-1) { + dum = 1.0/(a[j][j]); + for (i = j + 1; i < n; i++) a[i][j] *= dum; + } + } +} + diff --git a/src/MATRIX/pythag.cc b/src/MATRIX/pythag.cc new file mode 100644 index 0000000..2f6a440 --- /dev/null +++ b/src/MATRIX/pythag.cc @@ -0,0 +1,12 @@ +#include "JSC.h" +using namespace std; + +DP JSC::pythag(DP a, DP b) +{ + DP absa, absb; + + absa = fabs(a); + absb = fabs(b); + if (absa > absb) return absa * sqrt(1.0 + (absb * absb/(absa * absa))); + else return (absb == 0.0 ? 0.0 : absb * sqrt(1.0 + (absa * absa/(absb * absb)))); +} diff --git a/src/MATRIX/tqli.cc b/src/MATRIX/tqli.cc new file mode 100644 index 0000000..95076e9 --- /dev/null +++ b/src/MATRIX/tqli.cc @@ -0,0 +1,59 @@ +#include "JSC.h" + +using namespace std; + +void JSC::tqli(Vect_DP& d, Vect_DP& e, SQMat_DP& z) +{ + int m, l, iter, i, k; + DP s, r, p, g, f, dd, c, b; + + int n = d.size(); + for (i = 1; i < n; i++) e[i-1] = e[i]; + e[n-1] = 0.0; + for (l = 0; l < n; l++) { + iter = 0; + do { + for (m = l; m < n-1; m++) { + dd = fabs(d[m]) + fabs(d[m+1]); + if (fabs(e[m]) + dd == dd) break; + } + if (m != l) { + if (iter++ == 30) { + cout << "Too many iterations in tqli" << endl; + exit(1); + } + g = (d[l + 1] - d[l])/(2.0 * e[l]); + r = pythag(g, 1.0); + g = d[m] - d[l] + e[l]/(g + SIGN(r, g)); + s = c = 1.0; + p = 0.0; + for (i = m - 1; i >= l; i--) { + f = s * e[i]; + b = c * e[i]; + e[i + 1] = (r = pythag(f,g)); + if (r == 0.0) { + d[i + 1] -= p; + e[m] = 0.0; + break; + } + s = f/r; + c = g/r; + g = d[i + 1] - p; + r = (d[i] - g) * s + 2.0 * c * b; + d[i + 1] = g + (p = s * r); + g = c * r - b; + for (k = 0; k < n; k++) { + f = z[k][i + 1]; + z[k][i + 1] = s * z[k][i] + c * f; + z[k][i] = c * z[k][i] - s * f; + } + } + if (r == 0.0 && i >= l) continue; + d[l] -= p; + e[l] = g; + e[m] = 0.0; + } + } while (m != l); + } +} + diff --git a/src/MATRIX/tred2.cc b/src/MATRIX/tred2.cc new file mode 100644 index 0000000..f57f6b2 --- /dev/null +++ b/src/MATRIX/tred2.cc @@ -0,0 +1,63 @@ +#include "JSC.h" +using namespace std; + +void JSC::tred2 (SQMat_DP& a, Vect_DP& d, Vect_DP& e) +{ + int l, k, j, i; + DP scale, hh, h, g, f; + + int n = a.size(); + for (i = n-1; i > 0; i--) { + l = i - 1; + h = scale = 0.0; + if (l > 0) { + for (k = 0; k < l + 1; k++) scale += fabs(a[i][k]); + if (scale == 0.0) e[i] = a[i][l]; + else { + // scale = 1.0; // <- added this myself... + for (k = 0; k < l + 1; k++) { + a[i][k] /= scale; + h += a[i][k] * a[i][k]; + } + f = a[i][l]; + g = (f >= 0.0 ? -sqrt(h) : sqrt(h)); + e[i] = scale * g; + h -= f * g; + a[i][l] = f - g; + f = 0.0; + for (j = 0; j < l + 1; j++) { + a[j][i] = a[i][j]/h; + g = 0.0; + for (k = 0; k < j + 1; k++) g += a[j][k] * a[i][k]; + for (k = j + 1; k < l + 1; k++) g += a[k][j] * a[i][k]; + e[j] = g/h; + f += e[j] * a[i][j]; + } + hh = f/(h +h); + for (j = 0; j < l + 1; j++) { + f = a[i][j]; + e[j] = g = e[j] - hh * f; + for (k = 0; k < j + 1; k++) a[j][k] -= (f * e[k] + g * a[i][k]); + } + } + } else + e[i] = a[i][l]; + d[i] = h; + } + d[0] = 0.0; + e[0] = 0.0; + + for (i = 0; i < n; i++) { + l = i; + if (d[i] != 0.0) { + for (j = 0; j < l; j++) { + g = 0.0; + for (k = 0; k < l; k++) g += a[i][k] * a[k][j]; + for (k = 0; k < l; k++) a[k][j] -= g * a[k][i]; + } + } + d[i] = a[i][i]; + a[i][i] = 1.0; + for (j = 0; j < l; j++) a[j][i] = a[i][j] = 0.0; + } +} // void tred2 diff --git a/src/NRG/NRG_DME_Matrix_Block_builder.cc b/src/NRG/NRG_DME_Matrix_Block_builder.cc new file mode 100644 index 0000000..a080dfd --- /dev/null +++ b/src/NRG/NRG_DME_Matrix_Block_builder.cc @@ -0,0 +1,244 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: NRG_DFF_Matrix_Block_builder.cc + +Purpose: calculate matrix elements of density operator between + selected states in NRG for a selected block of the whole matrix. + For collaboration with Robert Konik. + +***********************************************************/ +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + + void Build_DME_Matrix_Block_for_NRG (DP c_int, DP L, int N, int iKmin, int iKmax, int Nstates_required, bool symmetric_states, int iKmod, + int weighing_option, int nrglabel_left_begin, int nrglabel_left_end, int nrglabel_right_begin, int nrglabel_right_end, + int block_option, DP* DME_block_1, DP* DME_block_2, Vect_DP Kweight) + { + // Given a list of states produced by Select_States_for_NRG, this function + // computes all the density form factors < lstate| \rho | rstate > + // for all id_left_begin <= id_left <= id_left_end (similarly for right states). + + // The block_option flag determines what kind of data is written in the 2 DME blocks. + // If !symmetric_states, then only block 1 is filled. + // If symmetric_states, then + // if block_option == 1, only block 1 is filled with the symmetric states ME. + // if block_option == 2, both blocks 1 and 2 are filled with ME and ME (after parity) respectively. + + // ASSUMPTIONS: + // We assume the DME_blocks are already reserved in memory. + // If !symmetric states, DME_block_2 doesn't need to be allocated. + + if (nrglabel_left_begin < 0 || nrglabel_right_begin < 0) JSCerror("beginning nrglabels negative in Build_DME_Matrix_Block_for_NRG"); + if (nrglabel_left_end >= Nstates_required) JSCerror("nrglabel_left_end too large in Build_DME_Matric_Block_for_NRG"); + if (nrglabel_right_end >= Nstates_required) JSCerror("nrglabel_right_end too large in Build_DME_Matric_Block_for_NRG"); + + if (nrglabel_left_begin > nrglabel_left_end) JSCerror("nrglabels of left states improperly chosen in DME block builder."); + if (nrglabel_right_begin > nrglabel_right_end) JSCerror("nrglabels of right states improperly chosen in DME block builder."); + // DME_block is a pointer of size row_l * col_l, where + int block_row_length = nrglabel_right_end - nrglabel_right_begin + 1; + int block_col_length = nrglabel_left_end - nrglabel_left_begin + 1; + + // iKmod is an integer specifying that we're only interested in momenta multiples of + // a base unit. In other words, if the perturbation potential repeats itself iKmod times + // on the periodic cycle on which the theory is defined, then we only need states such + // that iKl - iKr = (integer) * iKmod. We impose this constraint at the same time as + // the check on the weight of the state. + + // We start by reading off the list of selected states: + + stringstream NRG_stringstream; + string NRG_string; + + NRG_stringstream << "States_"; + NRG_stringstream << "c_" << c_int << "_L_" << L << "_N_" << N + << "_iKmin_" << iKmin << "_iKmax_" << iKmax << "_Nstates_" << Nstates_required + << "_Sym_" << symmetric_states << "_iKmod_" << iKmod << "_wopt_" << weighing_option << ".nrg"; + + NRG_string = NRG_stringstream.str(); + const char* NRG_Cstr = NRG_string.c_str(); + + ifstream infile; + infile.open(NRG_Cstr); + + if (infile.fail()) { + cout << NRG_Cstr << endl; + JSCerror("The input file was not opened successfully in Build_DME_Matrix_between_Selected_States_for_NRG. "); + } + + // Read the whole data file: + //const int MAXDATA = 5000000; + const int MAXDATA = JSC::max(nrglabel_left_end, nrglabel_right_end) + 10; // 10 for safety... + + int* nrglabel = new int[MAXDATA]; + DP* omega = new DP[MAXDATA]; + int* iK = new int[MAXDATA]; + string* label = new string[MAXDATA]; + bool* sym = new bool[MAXDATA]; + DP* weight = new DP[MAXDATA]; + + int Ndata = 0; + while (((infile.peek()) != EOF) && (Ndata < MAXDATA)) { + + infile >> nrglabel[Ndata]; + if (nrglabel[Ndata] != Ndata) JSCerror("reading states nrglabels wrong in NRG_DME_Matrix_Block_builder"); + infile >> omega[Ndata]; + infile >> iK[Ndata]; + infile >> label[Ndata]; + infile >> sym[Ndata]; + infile >> weight[Ndata]; + + Ndata++; + } + infile.close(); + + // We first reconstruct all the needed eigenstates: + + // Vector containing all the kept states: + Vect Kept_States_left(block_col_length); + Vect Kept_States_right(block_row_length); + + // Dummy state for calculations + LiebLin_Bethe_State Scanstate (c_int, L, N); + + // State list containing all the types of states + Scan_State_List List_of_Basic_States ('Z', Scanstate); + + for (int nrglabel_left = nrglabel_left_begin; nrglabel_left <= nrglabel_left_end; ++nrglabel_left) { + + Scanstate = List_of_Basic_States.Return_State (Extract_Base_Label(label[nrglabel_left])); + Scanstate.Set_to_Label (label[nrglabel_left]); + + Scanstate.Compute_All(true); + + Kept_States_left[nrglabel_left - nrglabel_left_begin] = Scanstate; + + if (!Scanstate.conv) cout << "State of label " << label[nrglabel_left] << " did not converge after " << Scanstate.iter_Newton + << " Newton step; diffsq = " << Scanstate.diffsq << endl; + } // for nrglabel_left + + for (int nrglabel_right = nrglabel_right_begin; nrglabel_right <= nrglabel_right_end; ++nrglabel_right) { + + Scanstate = List_of_Basic_States.Return_State (Extract_Base_Label(label[nrglabel_right])); + Scanstate.Set_to_Label (label[nrglabel_right]); + + Scanstate.Compute_All(true); + + Kept_States_right[nrglabel_right - nrglabel_right_begin] = Scanstate; + + if (!Scanstate.conv) cout << "State of label " << label[nrglabel_right] << " did not converge after " << Scanstate.iter_Newton + << " Newton step; diffsq = " << Scanstate.diffsq << endl; + } // for nrglabel_left + + + // Now that we have all the states, we can do the full calculation of all cross form factors: + + for (int ll = 0; ll < block_col_length; ++ll) { + for (int lr = 0; lr < block_row_length; ++lr) { + + if (Kept_States_left[ll].conv && Kept_States_right[lr].conv + //&& abs(Kept_States[il].iK - Kept_States[ir].iK) % iKmod == 0) + && abs(Kept_States_left[ll].iK) % iKmod == 0 + && abs(Kept_States_right[lr].iK) % iKmod == 0) { + + if (!symmetric_states) { + DME_block_1[ll* block_row_length + lr] = // By convention, put the lowest energy state to the left + Kweight[abs(Kept_States_left[ll].iK - Kept_States_right[lr].iK)] + * (Kept_States_left[ll].E <= Kept_States_right[lr].E + ? real(exp(ln_Density_ME (Kept_States_left[ll], Kept_States_right[lr]))) + : real(exp(ln_Density_ME (Kept_States_right[lr], Kept_States_left[ll])))); + // We don't do anything with block 2, we don't even assume it's been allocated. + } + + else if (symmetric_states) { + // Check parity of both states: + bool Lstate_parity_inv = Kept_States_left[ll].Check_Symmetry(); + bool Rstate_parity_inv = Kept_States_right[lr].Check_Symmetry(); + + if (Lstate_parity_inv && Rstate_parity_inv) { + DME_block_1[ll* block_row_length + lr] = Kweight[abs(Kept_States_left[ll].iK - Kept_States_right[lr].iK)] + * real(exp(ln_Density_ME (Kept_States_left[ll], Kept_States_right[lr]))); + if (block_option == 2) DME_block_2[ll* block_row_length + lr] = 0.0; + } + else if (!Lstate_parity_inv && !Rstate_parity_inv) { + LiebLin_Bethe_State PRstate = Kept_States_right[lr]; + PRstate.Parity_Flip(); + + if (block_option == 1) { + DME_block_1[ll* block_row_length + lr] = + (Kweight[abs(Kept_States_left[ll].iK - Kept_States_right[lr].iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], Kept_States_right[lr]))) + + Kweight[abs(Kept_States_left[ll].iK - PRstate.iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], PRstate)))); + // We don't do anything with block 2, we don't even assume it's been allocated. + } + else if (block_option == 2) { + DME_block_1[ll* block_row_length + lr] = + Kweight[abs(Kept_States_left[ll].iK - Kept_States_right[lr].iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], Kept_States_right[lr]))); + DME_block_2[ll* block_row_length + lr] = + Kweight[abs(Kept_States_left[ll].iK - PRstate.iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], PRstate))); + } + } + + else if (Lstate_parity_inv) { + LiebLin_Bethe_State PRstate = Kept_States_right[lr]; + PRstate.Parity_Flip(); + + if (block_option == 1) { + DME_block_1[ll* block_row_length + lr] = sqrt(0.5) * + (Kweight[abs(Kept_States_left[ll].iK - Kept_States_right[lr].iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], Kept_States_right[lr]))) + + Kweight[abs(Kept_States_left[ll].iK - PRstate.iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], PRstate)))); + // We don't do anything with block 2, we don't even assume it's been allocated. + } + else if (block_option == 2) { + DME_block_1[ll* block_row_length + lr] = sqrt(0.5) * + Kweight[abs(Kept_States_left[ll].iK - Kept_States_right[lr].iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], Kept_States_right[lr]))); + DME_block_2[ll* block_row_length + lr] = sqrt(0.5) * + Kweight[abs(Kept_States_left[ll].iK - PRstate.iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], PRstate))); + } + } + + else { // only R state is parity invariant, flip left: + LiebLin_Bethe_State PLstate = Kept_States_left[ll]; + PLstate.Parity_Flip(); + + if (block_option == 1) { + DME_block_1[ll* block_row_length + lr] = sqrt(0.5) * + (Kweight[abs(Kept_States_left[ll].iK - Kept_States_right[lr].iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], Kept_States_right[lr]))) + + Kweight[abs(PLstate.iK - Kept_States_right[lr].iK)] * real(exp(ln_Density_ME (PLstate, Kept_States_right[lr])))); + // We don't do anything with block 2, we don't even assume it's been allocated. + } + else if (block_option == 2) { + DME_block_1[ll* block_row_length + lr] = sqrt(0.5) * + Kweight[abs(Kept_States_left[ll].iK - Kept_States_right[lr].iK)] * real(exp(ln_Density_ME (Kept_States_left[ll], Kept_States_right[lr]))); + DME_block_2[ll* block_row_length + lr] = sqrt(0.5) * + Kweight[abs(PLstate.iK - Kept_States_right[lr].iK)] * real(exp(ln_Density_ME (PLstate, Kept_States_right[lr]))); + } + } + + } // if (symmetric_states) + } // if conv & iKmod condition fulfilled + + else { + DME_block_1[ll* block_row_length + lr] = 0.0; + if (symmetric_states && block_option == 2) DME_block_2[ll* block_row_length + lr] = 0.0; // condition, to prevent segfault if DME_block_2 not allocated. + } + + //cout << ll << "\t" << lr << "\t" << DME_block[ll* block_row_length + lr] << endl; + + } // for lr + } // for ll + + return; + } + + +} // namespace JSC diff --git a/src/NRG/NRG_K_Weight_integrand.cc b/src/NRG/NRG_K_Weight_integrand.cc new file mode 100644 index 0000000..b5f4ccd --- /dev/null +++ b/src/NRG/NRG_K_Weight_integrand.cc @@ -0,0 +1,47 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: NRG_K_Weight_integrand.cc + +Purpose: for NRG, this function specifies the momentum-dependent + part of the perturbation prefactor. + Collaboration with Robert Konik. + +***********************************************************/ +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + // This function defines the integrand for momentum-dependent perturbation prefactor + DP K_Weight_integrand (Vect_DP args) + { + // Momentum-dependent part of weight used in selecting kept form factors. + // This function gives the integrand, with the weight being + // \int_{-1/2}^{1/2} dx integrand(x) + + DP value; + + { // Case of a harmonic trap of strength omega_trap centered on 0: + + // args[0] == x + // args[1] == iK + // args[2] == L + // args[3] == omega_trap + + value = 12.0 * args[0] * args[0] * cos(twoPI * args[0] * args[1]) * pow(args[2], 2.0) * pow(args[3], 3.0); + } + + return(value); + } + + + +} // namespace JSC diff --git a/src/NRG/NRG_State_Selector.cc b/src/NRG/NRG_State_Selector.cc new file mode 100644 index 0000000..904f929 --- /dev/null +++ b/src/NRG/NRG_State_Selector.cc @@ -0,0 +1,461 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: NRG_State_Selector.cc + +Purpose: select states for numerical RG method, + collaboration with Robert Konik. + +***********************************************************/ +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + DP Estimate_Contribution_of_Single_ph_Annihilation_Path_to_2nd_Order_PT (LiebLin_Bethe_State& TopState, + LiebLin_Bethe_State& GroundState, + //Vect_DP Weight_integral) + Vect >& FT_of_potential) + { + //cout << TopState.label << "\t" << GroundState.label << endl; + //cout << TopState.lambdaoc << endl; + + DP contrib_estimate = 0.0; + + // Define OriginIx2 for labelling: + Vect OriginIx2 = GroundState.Ix2; + + // Calculate the number of particles in this state: + State_Label_Data topdata = Read_State_Label (TopState.label, OriginIx2); + int nphpairs = topdata.nexc[0]; + + if (nphpairs == 0) JSCerror("Trying to annihilate ground state in Estimate_Contribution..."); + + + DP densityME = real(exp(ln_Density_ME (GroundState, TopState))); + if (is_nan(densityME)) { + cout << "ME is nan: label = " << TopState.label << endl; + JSCerror("ME didn't return value."); + } + + int nr_cont = 0; + + // Add first-order PT contribution, and 2nd order from ground state (V_{00} == 1) + //contrib_estimate += Weight_integral[Weight_integral.size()/2 + (TopState.iK - GroundState.iK)] + //* (densityME/(GroundState.E - TopState.E)) * (1.0 - (GroundState.N/GroundState.L)/(GroundState.E - TopState.E)); + contrib_estimate += abs(FT_of_potential[FT_of_potential.size()/2 + (TopState.iK - GroundState.iK)] + * (densityME/(GroundState.E - TopState.E)) + * (1.0 - (GroundState.N/GroundState.L)/(GroundState.E - TopState.E))); + + nr_cont++; + + // Add second order PT contribution coming from TopState: + //contrib_estimate += Weight_integral[Weight_integral.size()/2 + 0] + //* Weight_integral[Weight_integral.size()/2 + (TopState.iK - GroundState.iK)] + //* 1.0 * densityME * (GroundState.N/GroundState.L) // 1.0 is V_{TopState, TopState} + ///((GroundState.E - TopState.E) * (GroundState.E - TopState.E)); + contrib_estimate += abs(FT_of_potential[FT_of_potential.size()/2 + 0] + * FT_of_potential[FT_of_potential.size()/2 + (TopState.iK - GroundState.iK)] + * 1.0 * densityME * (GroundState.N/GroundState.L) // 1.0 is V_{TopState, TopState} + /((GroundState.E - TopState.E) * (GroundState.E - TopState.E))); + + nr_cont++; + + //cout << "Here b" << endl; + + // Now add 2nd order terms coming from single particle-hole annihilation paths: + // this is only to be included for states with at least 4 excitations (2 ph pairs) + + if (nphpairs >= 2) { + + for (int ipart = 0; ipart < nphpairs; ++ipart) { + for (int ihole = 0; ihole < nphpairs; ++ihole) { + + LiebLin_Bethe_State DescendedState = TopState; + //cout << "Here 2a" << "\tipart " << ipart << "\tihole " << ihole << " out of " << nphpairs << " nphpairs, label " << TopState.label << endl; + //cout << "TopState.Ix2 " << TopState.Ix2 << endl; + //cout << "DescendedIx2 " << DescendedState.Ix2 << endl; + DescendedState.Annihilate_ph_pair(ipart, ihole, OriginIx2); + DescendedState.Compute_All(true); + //cout << "DescendedIx2 " << DescendedState.Ix2 << endl; + + DP densityME_top_desc = real(exp(ln_Density_ME (TopState, DescendedState))); + DP densityME_desc_ground = real(exp(ln_Density_ME (DescendedState, GroundState))); + + //contrib_estimate += Weight_integral[Weight_integral.size()/2 + (TopState.iK - DescendedState.iK)] + //* Weight_integral[Weight_integral.size()/2 + (DescendedState.iK - GroundState.iK)] + //* densityME_top_desc * densityME_desc_ground + ///((GroundState.E - TopState.E) * (GroundState.E - DescendedState.E)); + + // if intermediate state has momentum within allowable window, OK, otherwise discard contribution: + if (abs(TopState.iK - DescendedState.iK) < FT_of_potential.size()/2 && + abs(DescendedState.iK - GroundState.iK) < FT_of_potential.size()/2) { + + contrib_estimate += abs(FT_of_potential[FT_of_potential.size()/2 + (TopState.iK - DescendedState.iK)] + * FT_of_potential[FT_of_potential.size()/2 + (DescendedState.iK - GroundState.iK)] + * densityME_top_desc * densityME_desc_ground + /((GroundState.E - TopState.E) * (GroundState.E - DescendedState.E))); + + nr_cont++; + } + + if (nphpairs >= 3) { // go one step further + for (int ipart2 = ipart; ipart2 < nphpairs - 1; ++ipart2) { + for (int ihole2 = ihole; ihole2 < nphpairs - 1; ++ihole2) { + + LiebLin_Bethe_State DescendedState2 = DescendedState; + //cout << "Here 3a" << "\tipart2 " << ipart2 << "\tihole2 " << ihole2 << endl; + //cout << DescendedState2.Ix2 << endl; + DescendedState2.Annihilate_ph_pair(ipart2, ihole2, OriginIx2); + DescendedState2.Compute_All(true); + //cout << DescendedState2.Ix2 << endl; + //cout << DescendedState2.lambdaoc << endl; + + DP densityME_top_desc2 = real(exp(ln_Density_ME (TopState, DescendedState2))); + DP densityME_desc2_ground = real(exp(ln_Density_ME (DescendedState2, GroundState))); + + //contrib_estimate += Weight_integral[Weight_integral.size()/2 + (TopState.iK - DescendedState2.iK)] + //* Weight_integral[Weight_integral.size()/2 + (DescendedState2.iK - GroundState.iK)] + //* densityME_top_desc2 * densityME_desc2_ground + ///((GroundState.E - TopState.E) * (GroundState.E - DescendedState2.E)); + + // if intermediate state has momentum within allowable window, OK, otherwise discard contribution: + if (abs(TopState.iK - DescendedState2.iK) < FT_of_potential.size()/2 && + abs(DescendedState2.iK - GroundState.iK) < FT_of_potential.size()/2) { + + contrib_estimate += abs(FT_of_potential[FT_of_potential.size()/2 + (TopState.iK - DescendedState2.iK)] + * FT_of_potential[FT_of_potential.size()/2 + (DescendedState2.iK - GroundState.iK)] + * densityME_top_desc2 * densityME_desc2_ground + /((GroundState.E - TopState.E) * (GroundState.E - DescendedState2.E))); + + nr_cont++; + } + + if (nphpairs >= 4) { // go one step further + for (int ipart3 = ipart2; ipart3 < nphpairs - 2; ++ipart3) { + for (int ihole3 = ihole2; ihole3 < nphpairs - 2; ++ihole3) { + + LiebLin_Bethe_State DescendedState3 = DescendedState2; + //cout << "Here 4a" << "\tipart3 " << ipart3 << "\tihole3 " << ihole3 << endl; + DescendedState3.Annihilate_ph_pair(ipart3, ihole3, OriginIx2); + //cout << DescendedState3.Ix2 << endl; + DescendedState3.Compute_All(true); + //cout << DescendedState3.Ix2 << endl; + + DP densityME_top_desc3 = real(exp(ln_Density_ME (TopState, DescendedState3))); + DP densityME_desc3_ground = real(exp(ln_Density_ME (DescendedState3, GroundState))); + + //contrib_estimate += Weight_integral[Weight_integral.size()/2 + (TopState.iK - DescendedState3.iK)] + //* Weight_integral[Weight_integral.size()/2 + (DescendedState3.iK - GroundState.iK)] + //* densityME_top_desc3 * densityME_desc3_ground + ///((GroundState.E - TopState.E) * (GroundState.E - DescendedState3.E)); + + // if intermediate state has momentum within allowable window, OK, otherwise discard contribution: + if (abs(TopState.iK - DescendedState3.iK) < FT_of_potential.size()/2 && + abs(DescendedState3.iK - GroundState.iK) < FT_of_potential.size()/2) { + + contrib_estimate += abs(FT_of_potential[FT_of_potential.size()/2 + (TopState.iK - DescendedState3.iK)] + * FT_of_potential[FT_of_potential.size()/2 + (DescendedState3.iK - GroundState.iK)] + * densityME_top_desc3 * densityME_desc3_ground + /((GroundState.E - TopState.E) * (GroundState.E - DescendedState3.E))); + + nr_cont++; + } + + } // for ihole3 + } // for ipart3 + } // if (nphpairs >= 4) + } // for ihole2 + } // for ipart2 + } // if (nphpairs >= 3) + + } // for ihole + } // for ipart + } // if nphpairs >= 2 + + //cout << "Here b" << endl; + //cout << "type_id " << TopState.type_id << "\tid " << TopState.id << "\tcontrib_est = " << contrib_estimate << "\tnr_cont = " << nr_cont << endl; + + return(contrib_estimate); + } + + /* + DP Estimate_Contribution_of_Base_to_2nd_Order_PT (LiebLin_Bethe_State& ScanState, LiebLin_Bethe_State& GroundState, + Vect_DP Weight_integral, LiebLin_States_and_Density_ME_of_Base& StatesBase) + { + // This function calculates the second-order perturbation theory contribution to + // state ScanState coming from states of base defined in StatesBase. + + DP sum_contrib = 0.0; + + for (int i = 0; i <= StatesBase.maxid; ++i) + sum_contrib += Weight_integral[Weight_integral.size()/2 + (ScanState.iK - StatesBase.State[i].iK)] + * Weight_integral[Weight_integral.size()/2 + (StatesBase.State[i].iK - GroundState.iK)] + * real(exp(ln_Density_ME (ScanState, StatesBase.State[i]))) * StatesBase.densityME[i] + /((GroundState.E - StatesBase.State[i].E) * (GroundState.E - ScanState.E)); + // The strange vector size is so that iK == 0 corresponds to index size/2, as per convention + return(sum_contrib); + } + */ + + void Select_States_for_NRG (DP c_int, DP L, int N, int iKmin, int iKmax, int Nstates_required, bool symmetric_states, int iKmod, + //int weighing_option, DP (*weight_integrand_fn) (Vect_DP), Vect_DP& args_to_weight_integrand) + int weighing_option, Vect >& FT_of_potential) + { + // This function reads an existing partition function file and determines whether + // each state is to be included in NRG by applying an energy, momentum and form factor criterion. + + // The weighing function flag determines what kind of ordering is required: + // weighing_option == 0: ordering in energy + // weighing_option == 1: ordering according to perturbation theory in single p-h annihilation path + // weighing_option == 2: same as 1, but output of list is ordered in weight + + + stringstream filenameprefix; + Data_File_Name (filenameprefix, 'Z', c_int, L, N, iKmin, iKmax, 0.0, 0.0, ""); + string prefix = filenameprefix.str(); + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + + stringstream NRG_stringstream; string NRG_string; + NRG_stringstream << "States_c_" << c_int << "_L_" << L << "_N_" << N + << "_iKmin_" << iKmin << "_iKmax_" << iKmax << "_Nstates_" << Nstates_required + << "_Sym_" << symmetric_states << "_iKmod_" << iKmod << "_wopt_" << weighing_option << ".nrg"; + + + RAW_string = RAW_stringstream.str(); + const char* RAW_Cstr = RAW_string.c_str(); + + NRG_string = NRG_stringstream.str(); + const char* NRG_Cstr = NRG_string.c_str(); + + ifstream infile; + infile.open(RAW_Cstr); + + if (infile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("The input file was not opened successfully in Select_States_for_NRG. "); + } + + ofstream NRG_outfile; + + NRG_outfile.open(NRG_Cstr); + if (NRG_outfile.fail()) JSCerror("Could not open NRG_outfile... "); + + //NRG_outfile.setf(ios::scientific); + NRG_outfile.precision(16); + + + // Read the whole data file: + /* + // estimate its size: + struct stat statbuf; + stat (RAW_Cstr, &statbuf); + int filesize = statbuf.st_size; + // Determine the number of entries approximately + int entry_size = 1* sizeof(double) + 2*sizeof(int) + 3* sizeof(long long int); + + int estimate_nr_entries = filesize/entry_size; + */ + + // Count the number of entries in raw file: + int estimate_nr_entries = 0; + string line; + while (!infile.eof()) { + getline(infile, line); + estimate_nr_entries++; + } + + const int MAXDATA = estimate_nr_entries; + //cout << "estimate_nr_entries: " << estimate_nr_entries << endl; + + DP* E = new DP[MAXDATA]; + int* iK = new int[MAXDATA]; + //int* conv = new int[MAXDATA]; + string* label = new string[MAXDATA]; + bool* sym = new bool[MAXDATA]; + + int Ndata = 0; + + infile.close(); + infile.open(RAW_Cstr); + + while (((infile.peek()) != EOF) && (Ndata < MAXDATA)) { + infile >> E[Ndata]; + infile >> iK[Ndata]; + //infile >> conv[Ndata]; + infile >> label[Ndata]; + Ndata++; + } + + //cout << "input " << Ndata << " data lines." << endl; + infile.close(); + + // Define the ground state: + LiebLin_Bethe_State GroundState (c_int, L, N); + GroundState.Compute_All(true); + + // Define OriginIx2 for labelling: + Vect OriginIx2 = GroundState.Ix2; + + Scan_State_List ScanStateList ('d', GroundState); + + // Build the momentum-dependent weight integral matrix: + + // To cover negative and positive momenta (in case potential is not symmetric), + // we define the Weight_integral vector entry with index size/2 as corresponding to iK == 0. + + /* + // DEPRECATED 25/1/2012: from now on (ABACUS++T_8 onwards), use FT of potential as argument to this function. + Vect_DP Weight_integral (0.0, 4* (iKmax - iKmin)); // we give a large window + + //Vect_DP args(0.0, 2); + DP req_rel_prec = 1.0e-6; + DP req_abs_prec = 1.0e-6; + int max_nr_pts = 10000; + + for (int iK = -Weight_integral.size()/2; iK < Weight_integral.size()/2; ++iK) { + + args_to_weight_integrand[1] = DP(iK); + + Integral_result answer = Integrate_optimal (weight_integrand_fn, args_to_weight_integrand, + 0, 0.0, 0.5, req_rel_prec, req_abs_prec, max_nr_pts); + + Weight_integral[Weight_integral.size()/2 + iK] = answer.integ_est; + + } + */ + + // Calculate weight of states using selection criterion function + + //DP absdensityME = 0.0; + DP* weight = new DP[Ndata]; + + // For weighing using 2nd order PT, we only trace over 2 excitation states (1 p-h pair) + // Start by constructing these four classes once and for all: + + for (int i = 0; i < Ndata; ++i) { + + //cout << i << " out of " << Ndata << "\tlabel = " << label[i] << endl; + + if (abs(iK[i]) % iKmod != 0) { // if iK not a multiple of iKmod: give stupidly high weight. + weight[i] = 1.0e+100; + sym[i] = false; // doesn't matter + } + + else { + + // Construct the state again, so that the density ME can be calculated + LiebLin_Bethe_State ScanState = ScanStateList.Return_State(Extract_Base_Label(label[i])); + ScanState.Set_to_Label (label[i]); + if (weighing_option == 1 || weighing_option == 2) ScanState.Compute_All(true); + sym[i] = ScanState.Check_Symmetry(); + + //cout << "Setting state " << i << "\t to label " << label[i] << "\tsym: " << sym[i] << endl; + + // WEIGHING THE STATES: ********************************************** + + State_Label_Data currentdata = Read_State_Label (label[i], OriginIx2); + if (currentdata.nexc[0] == 0) weight[i] = 0.0; + + else if (symmetric_states && iK[i] < 0 || iK[i] < iKmin || iK[i] > iKmax) weight[i] = 1.0e+100; + + //else if (symmetric_states && iK[i] == 0 && !ScanState.Check_Symmetry()) { + else if (symmetric_states && iK[i] == 0 && !sym[i]) { + // This state is at zero momentum but not symmetric. we keep it only if + // the first non-symmetric pair of quantum numbers is right-weighted: + int icheck = 0; + while (ScanState.Ix2[N-1-icheck] == -ScanState.Ix2[icheck]) icheck++; + if (ScanState.Ix2[N-1-icheck] > -ScanState.Ix2[icheck]) { + if (weighing_option == 0) weight[i] = E[i]; + else if (weighing_option == 1 || weighing_option == 2) + weight[i] = 1.0/(1.0e-100 + fabs(Estimate_Contribution_of_Single_ph_Annihilation_Path_to_2nd_Order_PT + //(ScanState, GroundState, Weight_integral))); + (ScanState, GroundState, FT_of_potential))); + } + else weight[i] = 1.0e+100; + } + + else { + if (weighing_option == 0) weight[i] = E[i]; + else if (weighing_option == 1 || weighing_option == 2) + weight[i] = 1.0/(1.0e-100 + fabs(Estimate_Contribution_of_Single_ph_Annihilation_Path_to_2nd_Order_PT + //(ScanState, GroundState, Weight_integral))); + (ScanState, GroundState, FT_of_potential))); + } + //cout << i << " out of " << Ndata << "\tlabel = " << label[i] << "\tweight = " << weight[i] << endl; + } + } // for i + + // Now order the states in increasing weight + + int* index = new int[Ndata]; + for (int i = 0; i < Ndata; ++i) index[i] = i; + + QuickSort(weight, index, 0, Ndata - 1); + + + // Select states by increasing weight, with a max of Nstates_required entries + + DP* E_kept = new DP[Nstates_required]; + int* iK_kept = new int[Nstates_required]; + //int* conv_kept = new int[Nstates_required]; + string* label_kept = new string[Nstates_required]; + bool* sym_kept = new bool[Nstates_required]; + DP* weight_kept = new DP[Nstates_required]; + + // Copy selected states into new vectors: + + for (int i = 0; i < JSC::min(Ndata, Nstates_required); ++i) { + E_kept[i] = E[index[i] ]; + iK_kept[i] = iK[index[i] ]; + //conv_kept[i] = conv[index[i] ]; + label_kept[i] = label[index[i] ]; + sym_kept[i] = sym[index[i] ]; + weight_kept[i] = weight[i]; + } + + + // If needed, order selected states by increasing energy: + + int* index_kept = new int[Nstates_required]; + for (int i = 0; i < Nstates_required; ++i) index_kept[i] = i; + + if (weighing_option == 1) // only need to do this if energy ordering is chosen + QuickSort (E_kept, index_kept, 0, Nstates_required - 1); + + // Output selected states: + for (int i = 0; i < Nstates_required; ++i) { + if (i > 0) NRG_outfile << endl; + NRG_outfile << i << "\t" << E_kept[i] << "\t" << iK_kept[index_kept[i] ] + //<< "\t" << conv_kept[index_kept[i] ] + << "\t" << label_kept[index_kept[i] ] + << "\t" << sym_kept[index_kept[i] ] << "\t" << weight_kept[index_kept[i] ]; + } + + delete[] E; + delete[] iK; + //delete[] conv; + delete[] label; + delete[] sym; + delete[] E_kept; + delete[] iK_kept; + //delete[] conv_kept; + delete[] label_kept; + delete[] sym_kept; + delete[] weight; + + NRG_outfile.close(); + + return; + } + + +} // namespace JSC diff --git a/src/ODSLF/ODSLF.cc b/src/ODSLF/ODSLF.cc new file mode 100644 index 0000000..aca0852 --- /dev/null +++ b/src/ODSLF/ODSLF.cc @@ -0,0 +1,1616 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/1DSLF/1DSLF.cc + +Purpose: defines functions for 1d spinless fermion classes. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + //*************************************************************************************************** + + // Function definitions: class ODSLF_Base + + ODSLF_Base::ODSLF_Base () : Mdown(0), Nrap(Vect()), Nraptot(0), Ix2_infty(Vect()), Ix2_max(Vect()) {} + + ODSLF_Base::ODSLF_Base (const ODSLF_Base& RefBase) // copy constructor + : Mdown(RefBase.Mdown), Nrap(Vect(RefBase.Nrap.size())), Nraptot(RefBase.Nraptot), + Ix2_infty(Vect(RefBase.Nrap.size())), Ix2_max(Vect(RefBase.Nrap.size())), + id(RefBase.id) + { + for (int i = 0; i < Nrap.size(); ++i) { + Nrap[i] = RefBase.Nrap[i]; + Ix2_infty[i] = RefBase.Ix2_infty[i]; + Ix2_max[i] = RefBase.Ix2_max[i]; + } + } + + ODSLF_Base::ODSLF_Base (const Heis_Chain& RefChain, int M) + : Mdown(M), Nrap(Vect(RefChain.Nstrings)), Nraptot(0), Ix2_infty(Vect(RefChain.Nstrings)), Ix2_max(Vect(RefChain.Nstrings)) + { + for (int i = 0; i < RefChain.Nstrings; ++i) Nrap[i] = 0; + Nrap[0] = M; + + Nraptot = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Nraptot += Nrap[i]; + + // The id of this is zero by definition + id = 0LL; + + // Now compute the Ix2_infty numbers + + (*this).Compute_Ix2_limits(RefChain); + + } + + ODSLF_Base::ODSLF_Base (const Heis_Chain& RefChain, const Vect& Nrapidities) + : Mdown(0), Nrap(Nrapidities), Nraptot(0), Ix2_infty(Vect(RefChain.Nstrings)), Ix2_max(Vect(RefChain.Nstrings)), + id (0LL) + { + + // Check consistency of Nrapidities vector with RefChain + + //if (RefChain.Nstrings != Nrapidities.size()) cout << "error: Nstrings = " << RefChain.Nstrings << "\tNrap.size = " << Nrapidities.size() << endl; + + if (RefChain.Nstrings != Nrapidities.size()) JSCerror("Incompatible Nrapidities vector used in ODSLF_Base constructor."); + + int Mcheck = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Mcheck += RefChain.Str_L[i] * Nrap[i]; + Mdown = Mcheck; + + Nraptot = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Nraptot += Nrap[i]; + + // Compute id + id += Nrapidities[0]; + long long int factor = 100000LL; + for (int i = 1; i < RefChain.Nstrings; ++i) { + id += factor * Nrapidities[i]; + factor *= 100LL; + } + + // Now compute the Ix2_infty numbers + + (*this).Compute_Ix2_limits(RefChain); + + } + + ODSLF_Base::ODSLF_Base (const Heis_Chain& RefChain, long long int id_ref) + : Mdown(0), Nrap(Vect(RefChain.Nstrings)), Nraptot(0), Ix2_infty(Vect(RefChain.Nstrings)), Ix2_max(Vect(RefChain.Nstrings)), + id (id_ref) + { + // Build Nrapidities vector from id_ref + + long long int factor = pow_ulli (10LL, 2* RefChain.Nstrings + 1); + long long int id_eff = id_ref; + for (int i = 0; i < RefChain.Nstrings - 1; ++i) { + Nrap[RefChain.Nstrings - 1 - i] = id_eff/factor; + id_eff -= factor * Nrap[RefChain.Nstrings - 1 - i]; + factor /= 100LL; + } + Nrap[0] = id_eff; + + //id = id_ref; + + //cout << "In ODSLF_Base constructor: id_ref = " << id_ref << " and Nrapidities = " << Nrap << endl; + + // Check consistency of Nrapidities vector with RefChain + + //if (RefChain.Nstrings != Nrap.size()) JSCerror("Incompatible Nrapidities vector used in ODSLFBase constructor."); + + int Mcheck = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Mcheck += RefChain.Str_L[i] * Nrap[i]; + Mdown = Mcheck; + + Nraptot = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Nraptot += Nrap[i]; + + // Now compute the Ix2_infty numbers + + (*this).Compute_Ix2_limits(RefChain); + } + + + ODSLF_Base& ODSLF_Base::operator= (const ODSLF_Base& RefBase) + { + if (this != & RefBase) { + Mdown = RefBase.Mdown; + Nrap = RefBase.Nrap; + Nraptot = RefBase.Nraptot; + Ix2_infty = RefBase.Ix2_infty; + Ix2_max = RefBase.Ix2_max; + id = RefBase.id; + } + return(*this); + } + + bool ODSLF_Base::operator== (const ODSLF_Base& RefBase) + { + bool answer = (Nrap == RefBase.Nrap); + + return (answer); + } + + bool ODSLF_Base::operator!= (const ODSLF_Base& RefBase) + { + bool answer = (Nrap != RefBase.Nrap); + + return (answer); + } + + void ODSLF_Base::Compute_Ix2_limits (const Heis_Chain& RefChain) + { + + if ((RefChain.Delta > 0.0) && (RefChain.Delta < 1.0)) { + + // Compute the Ix2_infty numbers + + DP sum1 = 0.0; + DP sum2 = 0.0; + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + sum1 = 0.0; + + for (int k = 0; k < RefChain.Nstrings; ++k) { + + sum2 = 0.0; + + sum2 += (RefChain.Str_L[j] == RefChain.Str_L[k]) ? 0.0 : 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j] * RefChain.par[k]) + - 0.5 * fabs(RefChain.Str_L[j] - RefChain.Str_L[k]) * RefChain.anis)); + sum2 += 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j] * RefChain.par[k]) + - 0.5 * (RefChain.Str_L[j] + RefChain.Str_L[k]) * RefChain.anis)); + + for (int a = 1; a < JSC::min(RefChain.Str_L[j], RefChain.Str_L[k]); ++a) + sum2 += 2.0 * 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j] * RefChain.par[k]) + - 0.5 * (fabs(RefChain.Str_L[j] - RefChain.Str_L[k]) + 2.0*a) * RefChain.anis)); + + sum1 += (Nrap[k] - ((j == k) ? 1 : 0)) * sum2; + } + + Ix2_infty[j] = (1.0/PI) * fabs(RefChain.Nsites * 2.0 * atan(tan(0.25 * PI * (1.0 + RefChain.par[j]) + - 0.5 * RefChain.Str_L[j] * RefChain.anis)) - sum1); + + } // The Ix2_infty are now set. + + // Now compute the Ix2_max limits + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + Ix2_max[j] = int(floor(Ix2_infty[j])); // sets basic integer + + // Reject formally infinite rapidities (i.e. if Delta is root of unity) + + //cout << "Ix2_infty - Ix2_max = " << Ix2_infty[j] - Ix2_max[j] << endl; + //if (Ix2_infty[j] == Ix2_max[j]) { + //Ix2_max[j] -= 2; + //} + + // If Nrap is even, Ix2_max must be odd. If odd, then even. + //if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; DEACTIVATED FOR ODSLF ! + + // For ODSLF: parity of quantum numbers is complicated: + // if (N-1 + M_j + (M + 1) n_j + (1 - nu_j) N/2) is even, q# are integers, so Ix2_max must be even (conversely, if odd, then odd). + if ((Ix2_max[j] + RefChain.Nsites - 1 + Nrap[j] + (Nraptot + 1) * RefChain.Str_L[j] + (RefChain.par[j] == 1 ? 0 : RefChain.Nsites)) % 2) Ix2_max[j] -= 1; + + while (Ix2_max[j] > RefChain.Nsites) { + Ix2_max[j] -= 2; + } + } + } // if XXZ gapless + + else if (RefChain.Delta == 1.0) { + + // Compute the Ix2_infty numbers + + int sum1 = 0; + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + sum1 = 0; + + for (int k = 0; k < RefChain.Nstrings; ++k) { + + sum1 += Nrap[k] * (2 * JSC::min(RefChain.Str_L[j], RefChain.Str_L[k]) - ((j == k) ? 1 : 0)); + } + + //Ix2_infty[j] = (RefChain.Nsites - 1.0 + 2.0 * RefChain.Str_L[j] - sum1); + Ix2_infty[j] = (RefChain.Nsites + 1.0 - sum1); // to get counting right... + + } // The Ix2_infty are now set. + + // Now compute the Ix2_max limits + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + Ix2_max[j] = int(floor(Ix2_infty[j])); // sets basic integer + + // Give the correct parity to Ix2_max + + // If Nrap is even, Ix2_max must be odd. If odd, then even. + + //if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; DEACTIVATED FOR ODSLF ! + + // If Ix2_max equals Ix2_infty, we reduce it by 2: + + if (Ix2_max[j] == int(Ix2_infty[j])) Ix2_max[j] -= 2; + + while (Ix2_max[j] > RefChain.Nsites) { + Ix2_max[j] -= 2; + } + } + + } // if XXX AFM + + else if (RefChain.Delta > 1.0) { + + // Compute the Ix2_infty numbers + + int sum1 = 0; + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + sum1 = 0; + + for (int k = 0; k < RefChain.Nstrings; ++k) { + + sum1 += Nrap[k] * (2 * JSC::min(RefChain.Str_L[j], RefChain.Str_L[k]) - ((j == k) ? 1 : 0)); + } + + Ix2_infty[j] = (RefChain.Nsites - 1 + 2 * RefChain.Str_L[j] - sum1); + + } // The Ix2_infty are now set. + + // Now compute the Ix2_max limits + + for (int j = 0; j < RefChain.Nstrings; ++j) { + + Ix2_max[j] = int(floor(Ix2_infty[j])); // sets basic integer + + // Give the correct parity to Ix2_max + + // If Nrap is even, Ix2_max must be odd. If odd, then even. + + //if (!((Nrap[j] + Ix2_max[j]) % 2)) Ix2_max[j] -= 1; DEACTIVATED FOR ODSLF ! + + // If Ix2_max equals Ix2_infty, we reduce it by 2: + + //if (Ix2_max[j] == Ix2_infty[j]) Ix2_max[j] -= 2; + + while (Ix2_max[j] > RefChain.Nsites) { + Ix2_max[j] -= 2; + } + + // Fudge, for strings: + //if (RefChain.Str_L[j] >= 1) Ix2_max[j] += 2; + //Ix2_max[j] += 2; + } + + } // if XXZ_gpd + + } + + void ODSLF_Base::Scan_for_Possible_Types (Vect& possible_type_id, int& nfound, int base_level, Vect& Nexcitations) + { + if (base_level == 0) { + // We set all sectors 0, 1, 2, 3 + + int Nexc_max_0 = (Nrap[0] + 1)/2; // holes on R + int Nexc_max_1 = Nrap[0]/2; + int Nexc_max_2 = (Ix2_max[0] - Nrap[0] + 1)/2; + int Nexc_max_3 = Nexc_max_2; + + // Count the number of excitations in higher levels: + int Nexc_above = 0; + for (int i = 4; i < Nexcitations.size(); ++i) Nexc_above += Nexcitations[i]; + + for (int nphpairs = 0; nphpairs <= NEXC_MAX_HEIS/2 - Nexc_above; ++nphpairs) + for (int nexc0 = 0; nexc0 <= JSC::min(Nexc_max_0, nphpairs); ++nexc0) { + int nexc1 = nphpairs - nexc0; + for (int nexc2 = 0; nexc2 <= JSC::min(Nexc_max_2, nphpairs); ++nexc2) { + int nexc3 = nphpairs - nexc2; + + if (nexc1 >= 0 && nexc1 <= Nexc_max_1 && nexc3 >= 0 && nexc3 <= Nexc_max_3) { + // acceptable type found + Nexcitations[0] = nexc0; Nexcitations[1] = nexc1; Nexcitations[2] = nexc2; Nexcitations[3] = nexc3; + possible_type_id[nfound] = Ix2_Offsets_type_id (Nexcitations); + nfound++; + } + } + } + } + + else { // base_level > 0 + // We scan over the possible distributions of the excitations on the L and R sides: + + // Count the number of slots available on R: + //int Nexc_R_max = Ix2_max[base_level]/2 + 1; + int Nexc_R_max = (Ix2_max[base_level] + 2)/2; // bug in earlier line: if Ix2_max < 0, must have Nexc_R_max == 0, not 1. + int Nexc_L_max = (Ix2_max[base_level] + 1)/2; + for (int Nexcleft = 0; Nexcleft <= Nrap[base_level]; ++Nexcleft) + if (Nexcleft <= Nexc_L_max && Nrap[base_level] - Nexcleft <= Nexc_R_max) { + Nexcitations[2*base_level + 2] = Nrap[base_level] - Nexcleft; + Nexcitations[2*base_level + 3] = Nexcleft; + Scan_for_Possible_Types (possible_type_id, nfound, base_level - 1, Nexcitations); + } + } + + } + + Vect ODSLF_Base::Possible_Types () + { + // Given a base, this returns a vector of possible types + + Vect possible_type_id (0LL, 1000); + int nfound = 0; + + Vect Nexcitations (0, 2*Nrap.size() + 2); + + Scan_for_Possible_Types (possible_type_id, nfound, Nrap.size() - 1, Nexcitations); + + // Copy results into new vector: + Vect possible_type_id_found (nfound); + for (int i = 0; i < nfound; ++i) + possible_type_id_found[i] = possible_type_id[i]; + + return(possible_type_id_found); + } + + + //*************************************************************************************************** + + // Function definitions: class Ix2_Config + + ODSLF_Ix2_Config::ODSLF_Ix2_Config () {} + + ODSLF_Ix2_Config::ODSLF_Ix2_Config (const Heis_Chain& RefChain, int M) + : Nstrings(1), Nrap(Vect(M,1)), Nraptot(M), Ix2(new int*[1]) // single type of string here + //: Ix2 (Vect > (1)) + { + Ix2[0] = new int[M]; + //Ix2[0] = Vect (M); + + for (int j = 0; j < M; ++j) { + Ix2[0][j] = -(M - 1) + 2*j; + /* + cout << j << "\t" << Ix2[0][j] << endl; + cout << "Here 1" << endl; + cout << RefChain.Str_L[0] << endl; + // ODSLF: put back correct parity of quantum nr: + cout << "M = " << M << endl; + cout << Ix2[0][j] << endl; + cout << Nrap[0] << endl; + cout << Nraptot << endl; + cout << RefChain.Str_L[0] << endl; + cout << RefChain.par[0] << endl; + cout << RefChain.Nsites << endl; + if ((Ix2[0][j] + RefChain.Nsites - 1 + Nrap[0] + (Nraptot + 1) * RefChain.Str_L[0] + (RefChain.par[0] == 1 ? 0 : RefChain.Nsites)) % 2) Ix2[0][j] -= 1; + cout << "Here 2" << endl; + */ + // Use simplification: Nrap[0] = M = Nraptot, Str_L[0] = 1, par = 1, so + if ((Ix2[0][j] + RefChain.Nsites) % 2) Ix2[0][j] -= 1; + //OBSOLETE: 1 - (M % 2); // last term: correction for ODSLF + } + } + + ODSLF_Ix2_Config::ODSLF_Ix2_Config (const Heis_Chain& RefChain, const ODSLF_Base& base) + : Nstrings(RefChain.Nstrings), Nrap(base.Nrap), Nraptot(base.Nraptot), Ix2(new int*[RefChain.Nstrings]) + //: Ix2 (Vect > (RefChain.Nstrings)) + { + //Ix2[0] = new int[base.Mdown]; + Ix2[0] = new int[base.Nraptot]; + for (int i = 1; i < RefChain.Nstrings; ++i) Ix2[i] = Ix2[i-1] + base[i-1]; + + //for (int i = 0; i < RefChain.Nstrings; ++i) Ix2[i] = Vect (base[i]); + + // And now for the other string types... + for (int i = 0; i < RefChain.Nstrings; ++i) { + for (int j = 0; j < base[i]; ++j) Ix2[i][j] = -(base[i] - 1) + 2*j + 1 - (base[i] % 2); + } + + // ODSLF: put back correct parity of quantum nr: + for (int i = 0; i < RefChain.Nstrings; ++i) + for (int j = 0; j < base[i]; ++j) + if ((Ix2[i][j] + RefChain.Nsites - 1 + base.Nrap[i] + (base.Nraptot + 1) * RefChain.Str_L[i] + (RefChain.par[i] == 1 ? 0 : RefChain.Nsites)) % 2) + Ix2[i][j] -= 1; + + /* + // New meaning of negative parity: + for (int j = 1; j < base.Nrap.size(); ++j) + if (RefChain.par[j] == -1) // we `invert' the meaning of the quantum numbers + for (int alpha = 0; alpha < base[j]; ++alpha) { + if (Ix2[j][alpha] >= 0) Ix2[j][alpha] -= 2*(base.Ix2_max[j]/2 + 1); + else Ix2[j][alpha] += 2*(base.Ix2_max[j]/2 + 1); + } + */ + } + + ODSLF_Ix2_Config& ODSLF_Ix2_Config::operator= (const ODSLF_Ix2_Config& RefConfig) + { + if (this != &RefConfig) { + /* + Ix2 = Vect > (RefConfig.Ix2.size()); + for (int i = 0; i < RefConfig.Ix2.size(); ++i) { + Ix2[i] = Vect (RefConfig.Ix2[i].size()); + for (int j = 0; j < RefConfig.Ix2[i].size(); ++j) + Ix2[i][j] = RefConfig.Ix2[i][j]; + } + */ + + Nstrings = RefConfig.Nstrings; + Nrap = RefConfig.Nrap; + Nraptot = RefConfig.Nraptot; + if (Ix2 != 0) { + delete[] (Ix2[0]); + delete[] (Ix2); + } + Ix2 = new int*[Nstrings]; + Ix2[0] = new int[Nraptot]; + for (int i = 1; i < Nstrings; ++i) Ix2[i] = Ix2[i-1] + Nrap[i-1]; + for (int i = 0; i < Nstrings; ++i) + for (int j = 0; j < Nrap[i]; ++j) Ix2[i][j] = RefConfig.Ix2[i][j]; + + } + return(*this); + } + + ODSLF_Ix2_Config::~ODSLF_Ix2_Config() + { + if (Ix2 != 0) { + delete[] (Ix2[0]); + delete[] Ix2; + } + } + + std::ostream& operator<< (std::ostream& s, const ODSLF_Ix2_Config& RefConfig) + { + s << &RefConfig << "\t" << RefConfig.Nstrings << "\t" << RefConfig.Nraptot << endl; + for (int i = 0; i < RefConfig.Nstrings; ++i) s << RefConfig.Nrap[i] << "\t"; + s << endl; + for (int i = 0; i < RefConfig.Nstrings; ++i) { + s << "i " << i << "\t"; + for (int j = 0; j < RefConfig.Nrap[i]; ++j) s << RefConfig.Ix2[i][j] << "\t"; + s << endl; + } + return(s); + } + + //*************************************************************************************************** + + // Function definitions: class Lambda + + ODSLF_Lambda::ODSLF_Lambda () {} + + ODSLF_Lambda::ODSLF_Lambda (const Heis_Chain& RefChain, int M) + : Nstrings(1), Nrap(Vect(M,1)), Nraptot(M), lambda(new DP*[1]) // single type of string here + //: lambda(Vect > (1)) + { + lambda[0] = new DP[M]; + //lambda[0] = Vect (M); + + for (int j = 0; j < M; ++j) lambda[0][j] = 0.0; + } + + ODSLF_Lambda::ODSLF_Lambda (const Heis_Chain& RefChain, const ODSLF_Base& base) + : Nstrings(RefChain.Nstrings), Nrap(base.Nrap), Nraptot(base.Nraptot), lambda(new DP*[RefChain.Nstrings]) + //: lambda(Vect > (RefChain.Nstrings)) + { + //lambda[0] = new DP[base.Mdown]; + lambda[0] = new DP[base.Nraptot]; + for (int i = 1; i < RefChain.Nstrings; ++i) lambda[i] = lambda[i-1] + base[i-1]; + //for (int i = 0; i < RefChain.Nstrings; ++i) lambda[i] = Vect (base[i]); + + for (int i = 0; i < RefChain.Nstrings; ++i) { + for (int j = 0; j < base[i]; ++j) lambda[i][j] = 0.0; + } + } + + ODSLF_Lambda& ODSLF_Lambda::operator= (const ODSLF_Lambda& RefLambda) + { + if (this != &RefLambda) { + //lambda = RefLambda.lambda; + /* + lambda = Vect > (RefLambda.lambda.size()); + for (int i = 0; i < RefLambda.lambda.size(); ++i) { + lambda[i] = Vect (RefLambda.lambda[i].size()); + for (int j = 0; j < RefLambda.lambda[i].size(); ++j) lambda[i][j] = RefLambda.lambda[i][j]; + } + */ + + Nstrings = RefLambda.Nstrings; + Nrap = RefLambda.Nrap; + Nraptot = RefLambda.Nraptot; + if (lambda != 0) { + delete[] (lambda[0]); + delete[] (lambda); + } + lambda = new DP*[Nstrings]; + lambda[0] = new DP[Nraptot]; + for (int i = 1; i < Nstrings; ++i) lambda[i] = lambda[i-1] + Nrap[i-1]; + for (int i = 0; i < Nstrings; ++i) + for (int j = 0; j < Nrap[i]; ++j) lambda[i][j] = RefLambda.lambda[i][j]; + + } + return(*this); + } + + ODSLF_Lambda::~ODSLF_Lambda() + { + if (lambda != 0) { + delete[] (lambda[0]); + delete[] lambda; + } + } + + //*************************************************************************************************** + + // Function definitions: class ODSLF_Ix2_Offsets + + ODSLF_Ix2_Offsets::ODSLF_Ix2_Offsets () : base(), Tableau(Vect()), type_id(0LL), id(0LL), maxid(0LL) {}; + + ODSLF_Ix2_Offsets::ODSLF_Ix2_Offsets (const ODSLF_Ix2_Offsets& RefOffset) // copy constructor + : base(RefOffset.base), Tableau(Vect (2 * base.Nrap.size() + 2)), type_id(RefOffset.type_id), id(RefOffset.id), maxid(RefOffset.maxid) + { + for (int i = 0; i < 2 * base.Nrap.size() + 2; ++i) Tableau[i] = RefOffset.Tableau[i]; + } + + ODSLF_Ix2_Offsets::ODSLF_Ix2_Offsets (const ODSLF_Base& RefBase, long long int req_type_id) + // sets all tableaux to empty ones, with nparticles(req_type_id) at each level + { + // Build nparticles vector from req_type_id + + Vect nparticles(0, 2* RefBase.Nrap.size() + 2); + long long int factor = pow_ulli (10LL, nparticles.size() - 1); + long long int id_eff = req_type_id; + for (int i = 0; i < nparticles.size(); ++i) { + nparticles[nparticles.size() - 1 - i] = id_eff/factor; + id_eff -= factor * nparticles[nparticles.size() - 1 - i]; + factor /= 10LL; + } + + // Check if we've got the right vector... + long long int idcheck = Ix2_Offsets_type_id (nparticles); + if (idcheck != req_type_id) JSCerror("idcheck != req_type_id in Ix2_Offsets constructor."); + + (*this) = ODSLF_Ix2_Offsets(RefBase, nparticles); + } + + ODSLF_Ix2_Offsets::ODSLF_Ix2_Offsets (const ODSLF_Base& RefBase, Vect nparticles) // sets all tableaux to empty ones, with nparticles at each level + : base(RefBase), Tableau(Vect (2 * base.Nrap.size() + 2)), type_id(Ix2_Offsets_type_id (nparticles)), id(0LL), maxid(0LL) + { + + // Checks on nparticles vector: + + if (nparticles.size() != 2 * base.Nrap.size() + 2) JSCerror("Wrong nparticles.size in Ix2_Offsets constructor."); + + //if (base.Nrap[0] != (nparticles[3] + nparticles[2] + base.Mdown - nparticles[0] - nparticles[1])) JSCerror("Wrong Nrap[0] in Ix2_Offsets constructor."); + if (nparticles[3] + nparticles[2] != nparticles[0] + nparticles[1]) { + cout << nparticles[0] << "\t" << nparticles[1] << "\t" << nparticles[2] << "\t" << nparticles[3] << endl; + JSCerror("Wrong Npar[0-3] in Ix2_Offsets constructor."); + } + + for (int base_level = 1; base_level < base.Nrap.size(); ++ base_level) + if (base.Nrap[base_level] != nparticles[2*base_level + 2] + nparticles[2*base_level + 3]) { + cout << base_level << "\t" << base.Nrap[base_level] << "\t" << nparticles[2*base_level + 2] << "\t" << nparticles[2*base_level + 3] << endl; + JSCerror("Wrong Nrap[] in Ix2_Offsets constructor."); + } + + // nparticles[0,1]: number of holes on R and L side in GS interval + if (nparticles[0] > (base.Nrap[0] + 1)/2) JSCerror("nparticles[0] too large in Ix2_Offsets constructor."); + if (nparticles[1] > base.Nrap[0]/2) JSCerror("nparticles[1] too large in Ix2_Offsets constructor."); + + // nparticles[2,3]: number of particles of type 0 on R and L side out of GS interval + if (nparticles[2] > (base.Ix2_max[0] - base.Nrap[0] + 1)/2) JSCerror("nparticles[2] too large in Ix2_Offsets constructor."); + if (nparticles[3] > (base.Ix2_max[0] - base.Nrap[0] + 1)/2) JSCerror("nparticles[3] too large in Ix2_Offsets constructor."); + + for (int base_level = 1; base_level < base.Nrap.size(); ++ base_level) + //if ((nparticles[2*base_level + 2] > 0 && nparticles[2*base_level + 2] > (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2) + if ((nparticles[2*base_level + 2] > 0 && nparticles[2*base_level + 2] > (base.Ix2_max[base_level] + 2)/2) // ODSLF modif + //|| (nparticles[2*base_level + 3] > 0 && nparticles[2*base_level + 3] > (base.Ix2_max[base_level] - (base.Nrap[base_level] % 2) - 1)/2)) { + || (nparticles[2*base_level + 3] > 0 + //&& nparticles[2*base_level + 3] > base.Ix2_max[base_level] + 1 - (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2)) + && nparticles[2*base_level + 3] > (base.Ix2_max[base_level] + 1)/2)) // ODSLF modif + { + cout << base_level << "\t" << base.Ix2_max[base_level] << "\t" << base.Ix2_infty[base_level] << "\t" << nparticles[2*base_level + 2] << "\t" << (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2 + << "\t" << nparticles[2*base_level + 3] << "\t" << (base.Ix2_max[base_level] - (base.Nrap[base_level] % 2) - 1)/2 + << "\t" << (nparticles[2*base_level + 2] > 0) << "\t" << (nparticles[2*base_level + 2] > (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2) + //<< "\t" << (nparticles[2*base_level + 3] > 0) << "\t" << (nparticles[2*base_level + 3] > (base.Ix2_max[base_level] - (base.Nrap[base_level] % 2) - 1)/2) + << "\t" << (nparticles[2*base_level + 3] > 0) << "\t" + << (nparticles[2*base_level + 3] > base.Ix2_max[base_level] + 1 - (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2) + 2)/2) + << endl; + JSCerror("nparticles too large in Ix2_Offsets constructor."); + } + + // Check sum of rapidities + + // Holes in GS interval + Tableau[0] = Young_Tableau(nparticles[0], (base.Nrap[0] + 1)/2 - nparticles[0]); + Tableau[1] = Young_Tableau(nparticles[1], base.Nrap[0]/2 - nparticles[1], Tableau[0]); + + // Particles of type 0 out of GS interval + Tableau[2] = Young_Tableau(nparticles[2], (base.Ix2_max[0] - base.Nrap[0] + 1)/2 - nparticles[2], Tableau[0]); + Tableau[3] = Young_Tableau(nparticles[3], (base.Ix2_max[0] - base.Nrap[0] + 1)/2 - nparticles[3], Tableau[2]); + + // Tableaux of index i = 2,...: data about string type i/2-1. + for (int base_level = 1; base_level < base.Nrap.size(); ++base_level) { + Tableau[2*base_level + 2] = Young_Tableau(nparticles[2*base_level + 2], + //(base.Ix2_max[base_level] - ((base.Nrap[base_level]) % 2) + 2)/2 - nparticles[2*base_level + 2], Tableau[2]); + //(base.Ix2_max[base_level] - base.Nrap[base_level] % 2 + 2)/2 - nparticles[2*base_level + 2], Tableau[2]); + (base.Ix2_max[base_level] - ((base.Nrap[base_level] + 1) % 2))/2 + 1 - nparticles[2*base_level + 2], Tableau[2]); + Tableau[2*base_level + 3] = Young_Tableau(nparticles[2*base_level + 3], + //(base.Ix2_max[base_level] - base.Nrap[base_level] % 2)/2 - nparticles[2*base_level + 3], Tableau[3]); + (base.Ix2_max[base_level] - (base.Nrap[base_level] % 2) - 1)/2 + 1 - nparticles[2*base_level + 3], Tableau[3]); + } + + maxid = 1LL; + //id = Tableau[0].id; + for (int i = 0; i < nparticles.size(); ++i) { + maxid *= Tableau[i].maxid + 1LL; + //id += maxid + Tableau[i].id; + } + maxid -= 1LL; + + } + + ODSLF_Ix2_Offsets& ODSLF_Ix2_Offsets::operator= (const ODSLF_Ix2_Offsets& RefOffset) + { + if (this != &RefOffset) { + base = RefOffset.base; + Tableau = RefOffset.Tableau; + type_id = RefOffset.type_id; + id = RefOffset.id; + maxid = RefOffset.maxid; + } + return(*this); + } + + bool ODSLF_Ix2_Offsets::operator<= (const ODSLF_Ix2_Offsets& RefOffsets) + { + // Check whether all nonzero tableau row lengths in RefOffsets + // are <= than those in *this + + bool answer = true; + for (int level = 0; level < 4; ++level) { // check fundamental level only + //for (int level = 0; level < 2 * base.Nrap.size() + 2; ++level) { + // First check whether all rows which exist in both tableaux satisfy rule: + for (int tableau_level = 0; tableau_level < JSC::min(Tableau[level].Nrows, RefOffsets.Tableau[level].Nrows); ++tableau_level) + if (Tableau[level].Row_L[tableau_level] > RefOffsets.Tableau[level].Row_L[tableau_level]) + answer = false; + // Now check whether there exist extra rows violating rule: + for (int tableau_level = JSC::min(Tableau[level].Nrows, RefOffsets.Tableau[level].Nrows); tableau_level < Tableau[level].Nrows; ++tableau_level) + if (Tableau[level].Row_L[tableau_level] > 0) answer = false; + } + + return(answer); + } + + bool ODSLF_Ix2_Offsets::operator>= (const ODSLF_Ix2_Offsets& RefOffsets) + { + // Check whether all nonzero tableau row lengths in RefOffsets + // are >= than those in *this + + bool answer = true; + for (int level = 0; level < 4; ++level) { // check fundamental level only + //for (int level = 0; level < 2 * base.Nrap.size() + 2; ++level) { + // First check whether all rows which exist in both tableaux satisfy rule: + for (int tableau_level = 0; tableau_level < JSC::min(Tableau[level].Nrows, RefOffsets.Tableau[level].Nrows); ++tableau_level) + if (Tableau[level].Row_L[tableau_level] < RefOffsets.Tableau[level].Row_L[tableau_level]) + answer = false; + // Now check whether there exist extra rows violating rule: + for (int tableau_level = JSC::min(Tableau[level].Nrows, RefOffsets.Tableau[level].Nrows); tableau_level < RefOffsets.Tableau[level].Nrows; ++tableau_level) + if (RefOffsets.Tableau[level].Row_L[tableau_level] > 0) answer = false; + } + + return(answer); + } + + void ODSLF_Ix2_Offsets::Set_to_id (long long int idnr) + { + // The idnr of the Offset is given by + // sub_id[0] + (total number of tableaux of type 0) * (sub_id[1] + (total number of tableaux of type 1) * (sub_id[2] + ... + // + total number of tableaux of type (2*base.Nrap.size()) * sub_id[2*base.Nrap.size() + 1] + + //cout << "Called Ix2_Offsets::Set_to_id" << endl; + + if (idnr > maxid) { + cout << idnr << "\t" << maxid << endl; + JSCerror("idnr too large in offsets.Set_to_id."); + } + + id = idnr; + + Vect sub_id(0LL, 2*base.Nrap.size() + 2); + + long long int idnr_eff = idnr; + long long int temp_prod = 1LL; + + Vect result_choose(2*base.Nrap.size() + 2); + + for (int i = 0; i <= 2*base.Nrap.size(); ++i) { + //result_choose[i] = choose_lli(Tableau[i].Nrows + Tableau[i].Ncols, Tableau[i].Nrows); + result_choose[i] = Tableau[i].maxid + 1LL; + temp_prod *= result_choose[i]; + } + + for (int i = 2*base.Nrap.size() + 1; i > 0; --i) { + sub_id[i] = idnr_eff/temp_prod; + idnr_eff -= sub_id[i] * temp_prod; + temp_prod /= result_choose[i-1]; + } + sub_id[0] = idnr_eff; // what's left goes to the bottom... + + for (int i = 0; i <= 2*base.Nrap.size() + 1; ++i) { + //cout << "level = " << i << " Tableau.id = " << sub_id[i] << endl; + if ((Tableau[i].Nrows * Tableau[i].Ncols == 0) && (sub_id[i] != 0)) JSCerror("index too large in offset.Set_to_id."); + if (Tableau[i].id != sub_id[i]) Tableau[i].Set_to_id(sub_id[i]); + } + + Compute_type_id (); + + return; + } + + void ODSLF_Ix2_Offsets::Compute_type_id () + { + type_id = 0LL; + + for (int i = 0; i < 2*base.Nrap.size() + 2; ++i) { + Tableau[i].Compute_id(); + type_id += Tableau[i].Nrows * pow_ulli(10LL, i); + } + } + + void ODSLF_Ix2_Offsets::Compute_id () + { + long long int prod_maxid = 1LL; + + id = 0LL; + + for (int i = 0; i < 2*base.Nrap.size() + 2; ++i) { + Tableau[i].Compute_id(); + id += Tableau[i].id * prod_maxid; + prod_maxid *= Tableau[i].maxid + 1LL; + } + + } + + bool ODSLF_Ix2_Offsets::Add_Boxes_From_Lowest (int Nboxes, bool odd_sectors) + { + // Tries to add Nboxes to Young tableau in given sector parity. + // If Nboxes is too large to be accommodated, `false' is returned. + + ODSLF_Ix2_Offsets offsets_init(*this); // keep a copy in case it fails + + if (Nboxes < 0) JSCerror("Can't use Nboxes < 0 in Add_Boxes_From_Lowest"); + else if (Nboxes == 0) return(true); // nothing to do + + //cout << "Requesting Nboxes " << Nboxes << " in " << odd_sectors << " sectors." << endl; + + int Nboxes_put = 0; + int working_sector = odd_sectors; + // We start from the bottom up. + while (Nboxes_put < Nboxes && working_sector < 2*base.Nrap.size() + 2) { + //cout << "working sector " << working_sector << "\tNboxes_put " << Nboxes_put << endl; + //cout << "Tableau before put: " << Tableau[working_sector] << endl; + Nboxes_put += Tableau[working_sector].Add_Boxes_From_Lowest(Nboxes - Nboxes_put); + //cout << "Tableau after put: " << Tableau[working_sector] << endl; + working_sector += 2; + } + Compute_id(); + + if (Nboxes_put < Nboxes) (*this) = offsets_init; // reset ! + else if (Nboxes_put > Nboxes) { + cout << Nboxes_put << "\t" << Nboxes << endl; + JSCerror("Putting too many boxes in Ix2_Offsets::Add_Boxes_From_Lowest."); + } + return(Nboxes_put == Nboxes); + } + + //*************************************************************************************************** + + // Function definitions: class ODSLF_Ix2_Offsets_List + + ODSLF_Ix2_Offsets_List::ODSLF_Ix2_Offsets_List() : ndef(0), Offsets(Vect(NBASESMAX)) {}; + + ODSLF_Ix2_Offsets& ODSLF_Ix2_Offsets_List::Return_Offsets (ODSLF_Base& RefBase, Vect nparticles) + { + + long long int Ix2_Offsets_type_id_ref = Ix2_Offsets_type_id (nparticles); + int n = 0; + + while (n < ndef && !(Ix2_Offsets_type_id_ref == Offsets[n].type_id && RefBase == Offsets[n].base)) n++; + + if (n == ndef) { + Offsets[n] = ODSLF_Ix2_Offsets (RefBase, nparticles); + ndef++; + } + + if (ndef >= NBASESMAX) JSCerror("Ix2_Offsets_List: need bigger Offsets vector."); + + return(Offsets[n]); + } + + ODSLF_Ix2_Offsets& ODSLF_Ix2_Offsets_List::Return_Offsets (ODSLF_Base& RefBase, long long int req_type_id) + { + + int n = 0; + + while (n < ndef && !(req_type_id == Offsets[n].type_id && RefBase == Offsets[n].base)) n++; + + if (n == ndef) { + Offsets[n] = ODSLF_Ix2_Offsets (RefBase, req_type_id); + ndef++; + } + + if (ndef >= NBASESMAX) JSCerror("Ix2_Offsets_List: need bigger Offsets vector."); + + return(Offsets[n]); + } + + + + //*************************************************************************************************** + + // Function definitions: class ODSLF_Bethe_State + + ODSLF_Bethe_State::ODSLF_Bethe_State () + : chain(Heis_Chain()), base(ODSLF_Base()), offsets(ODSLF_Ix2_Offsets()), Ix2(ODSLF_Ix2_Config(chain, 1)), + lambda(ODSLF_Lambda(chain, 1)), BE(ODSLF_Lambda(chain, 1)), diffsq(0.0), conv(0), iter(0), iter_Newton(0), + E(0.0), iK(0), K(0.0), lnnorm(-100.0), base_id(0LL), type_id(0LL), id(0LL), maxid(0LL), nparticles(0) + { + }; + + ODSLF_Bethe_State::ODSLF_Bethe_State (const ODSLF_Bethe_State& RefState) // copy constructor + //: chain(RefState.chain), base(RefState.base), offsets(RefState.offsets), Ix2(Ix2_Config(RefState.chain, RefState.base.Mdown)), + // lambda(Lambda(RefState.chain, RefState.base.Mdown)), BE(Lambda(RefState.chain, RefState.base.Mdown)), + : chain(RefState.chain), base(RefState.base), offsets(RefState.offsets), Ix2(ODSLF_Ix2_Config(RefState.chain, RefState.base)), + lambda(ODSLF_Lambda(RefState.chain, RefState.base)), BE(ODSLF_Lambda(RefState.chain, RefState.base)), + diffsq(RefState.diffsq), conv(RefState.conv), iter(RefState.iter), iter_Newton(RefState.iter_Newton), + E(RefState.E), iK(RefState.iK), K(RefState.K), lnnorm(RefState.lnnorm), + //id(RefState.id), maxid(RefState.maxid) + base_id(RefState.base_id), type_id(RefState.type_id), id(RefState.id), maxid(RefState.maxid), + nparticles(RefState.nparticles) + { + // copy arrays into new ones + + /* + cout << "Here in Heis constructor state" << endl; + cout << "lambda " << lambda[0][0] << endl; + cout << "lambda OK" << endl; + + cout << "RefConfig: " << endl << RefState.Ix2 << endl; + cout << "(*this).Ix2: " << endl << Ix2 << endl; + */ + for (int j = 0; j < RefState.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < RefState.base[j]; ++j) { + Ix2[j][alpha] = RefState.Ix2[j][alpha]; + lambda[j][alpha] = RefState.lambda[j][alpha]; + } + } + //cout << "Heis constructor state OK" << endl; + } + + ODSLF_Bethe_State::ODSLF_Bethe_State (const ODSLF_Bethe_State& RefState, long long int type_id_ref) + : chain(RefState.chain), base(RefState.base), offsets(ODSLF_Ix2_Offsets(RefState.base, type_id_ref)), + Ix2(ODSLF_Ix2_Config(RefState.chain, RefState.base)), lambda(ODSLF_Lambda(RefState.chain, RefState.base)), + BE(ODSLF_Lambda(RefState.chain, RefState.base)), diffsq(1.0), conv(0), iter(0), iter_Newton(0), + E(0.0), iK(0), K(0.0), lnnorm(-100.0), + //id(0LL), maxid(0LL) + base_id(RefState.base_id), type_id(0LL), id(0LL), maxid(offsets.maxid), nparticles(0) + { + //cout << "Here in Heis constructor state type_id" << endl; + //cout << "lambda " << lambda[0][0] << endl; + //cout << "lambda OK" << endl; + + (*this).Set_Ix2_Offsets (offsets); + } + + ODSLF_Bethe_State::ODSLF_Bethe_State (const Heis_Chain& RefChain, int M) + : chain(RefChain), base(RefChain, M), offsets(base, 0LL), Ix2(ODSLF_Ix2_Config(RefChain, M)), lambda(ODSLF_Lambda(RefChain, M)), + BE(ODSLF_Lambda(RefChain, M)), diffsq(1.0), conv(0), iter(0), iter_Newton(0), + E(0.0), iK(0), K(0.0), lnnorm(-100.0), + //id(0LL), maxid(0LL) + base_id(0LL), type_id(0LL), id(0LL), maxid(offsets.maxid), nparticles(0) + { + //cout << "Here in Heis constructor chain M" << endl; + //cout << "requested M = " << M << endl; + //cout << "lambda " << lambda[0][0] << endl; + //cout << "lambda OK" << endl; + } + + ODSLF_Bethe_State::ODSLF_Bethe_State (const Heis_Chain& RefChain, const ODSLF_Base& RefBase) + : chain(RefChain), base(RefBase), offsets(RefBase, 0LL), Ix2(ODSLF_Ix2_Config(RefChain, RefBase)), lambda(ODSLF_Lambda(RefChain, RefBase)), + BE(ODSLF_Lambda(RefChain, RefBase)), diffsq(1.0), conv(0), iter(0), iter_Newton(0), E(0.0), iK(0), K(0.0), lnnorm(-100.0), + //id(0LL), maxid(0LL) + base_id(RefBase.id), type_id(0LL), id(0LL), maxid(offsets.maxid), nparticles(0) + { + // Check that the number of rapidities is consistent with Mdown + + //cout << "Here in Heis constructor chain base" << endl; + //cout << "lambda " << lambda[0][0] << endl; + //cout << "lambda OK" << endl; + + int Mcheck = 0; + for (int i = 0; i < RefChain.Nstrings; ++i) Mcheck += base[i] * RefChain.Str_L[i]; + if (RefBase.Mdown != Mcheck) JSCerror("Wrong M from Nrapidities input vector, in ODSLF_Bethe_State constructor."); + + } + + ODSLF_Bethe_State::ODSLF_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref) + : chain(RefChain), base(ODSLF_Base(RefChain, base_id_ref)), offsets(base, type_id_ref), Ix2(ODSLF_Ix2_Config(RefChain, base)), lambda(ODSLF_Lambda(RefChain, base)), + BE(ODSLF_Lambda(RefChain, base)), diffsq(1.0), conv(0), iter(0), iter_Newton(0), E(0.0), iK(0), K(0.0), lnnorm(-100.0), + //id(0LL), maxid(0LL) + base_id(base_id_ref), type_id(type_id_ref), id(0LL), maxid(offsets.maxid), nparticles(0) + { + //cout << "Here in Heis constructor chain base_id type_id" << endl; + //cout << "lambda " << lambda[0][0] << endl; + //cout << "lambda OK" << endl; + } + + void ODSLF_Bethe_State::Set_Ix2_Offsets (const ODSLF_Ix2_Offsets& RefOffset) // sets the Ix2 to given offsets + { + if (base != RefOffset.base) JSCerror("Offset with incompatible base used in ODSLF_Bethe_State::Set_Ix2_Offsets."); + /* + // Check if total number of holes agrees with number of rapidities in higher particles + int total_nr_holes = RefOffset.Tableau[0].Nrows + RefOffset.Tableau[1].Nrows; + int total_nr_rap_in_particles = 0; + for (int offsets_level = 2; offsets_level < RefOffset.Tableau.size(); offsets_level += 2) + total_nr_rap_in_particles += chain.Str_L[offsets_level/2 - 1] * (RefOffset.Tableau[offsets_level].Nrows + RefOffset.Tableau[offsets_level + 1].Nrows); + + cout << total_nr_holes << "\t" << total_nr_rap_in_particles << endl; + + if (total_nr_holes != total_nr_rap_in_particles) JSCerror("Offset with incompatible Nparticles used in ODSLF_Bethe_State::Set_Ix2_Offsets."); + */ + + // For each base_level, we make sure that the Ix2 are ordered: Ix2[0][j] < Ix2[0][k] if j < k. + + // Level 2: particles in R part outside GS interval + for (int j = 0; j < RefOffset.Tableau[2].Nrows; ++j) { + Ix2[0][base.Nrap[0] - 1 - j] + = base.Nrap[0] - 1 + 2* RefOffset.Tableau[2].Nrows - 2*j + 2*RefOffset.Tableau[2].Row_L[j]; + //cout << "R pcles, j = " << base.Nrap[0] - 1 - j << " Ix2[j] = " << Ix2[0][base.Nrap[0] - 1 - j] << endl; + //cout << "j = " << j << "\tbase.Nrap[0] " << base.Nrap[0] << "\t2*RefOffset.Tableau[2].Nrows " << 2*RefOffset.Tableau[2].Nrows << "\t2*RefOffset.Tableau[2].Row_L[j] " << 2*RefOffset.Tableau[2].Row_L[j] << endl; + //cout << base.Nrap[0] - 1 - j << "\t" << base.Nrap[0] - 1 + 2* RefOffset.Tableau[2].Nrows - 2*j + 2*RefOffset.Tableau[2].Row_L[j] << "\t" << Ix2[0][base.Nrap[0] - 1 - j] << "\t" << Ix2[0][9] << endl; + } + nparticles = RefOffset.Tableau[2].Nrows; + //cout << "Here a" << endl; + + // Level 0: holes in R part of GS interval + for (int j = 0; j < RefOffset.Tableau[0].Ncols; ++j) { + //Ix2[0][base.Nrap[0] - 1 - RefOffset.Tableau[2].Nrows - j] = base.Nrap[0] - 1 - 2*RefOffset.Tableau[0].Nrows - 2*j + 2*RefOffset.Tableau[0].Col_L[j]; + Ix2[0][base.Nrap[0] - 1 - RefOffset.Tableau[2].Nrows - j] = (base.Nrap[0] + 1) % 2 + 2*RefOffset.Tableau[0].Ncols - 2 - 2*j + 2*RefOffset.Tableau[0].Col_L[j]; + //cout << "R holes, j = " << base.Nrap[0] - 1 - RefOffset.Tableau[2].Nrows - j << " Ix2[j] = " << Ix2[0][base.Nrap[0] - 1 - RefOffset.Tableau[2].Nrows - j] << endl; + //cout << base.Nrap[0] << "\t" << (base.Nrap[0] + 1) % 2 << "\t" << RefOffset.Tableau[0].Ncols << "\t" << RefOffset.Tableau[0].Col_L[j] << endl; + } + nparticles += RefOffset.Tableau[0].Ncols; + + //cout << "Here b" << endl; + + // Level 1: holes in L part of GS interval + for (int j = 0; j < RefOffset.Tableau[1].Ncols; ++j) { + Ix2[0][base.Nrap[0] - 1 - RefOffset.Tableau[0].Ncols - RefOffset.Tableau[2].Nrows - j] + = -1 - (base.Nrap[0] % 2) - 2* j - 2*RefOffset.Tableau[1].Col_L[RefOffset.Tableau[1].Ncols - 1 - j]; + //cout << "L holes, j = " << base.Nrap[0] - 1 - RefOffset.Tableau[0].Ncols - RefOffset.Tableau[2].Nrows- j << " Ix2[j] = " << Ix2[0][base.Nrap[0] - 1 - RefOffset.Tableau[0].Ncols - RefOffset.Tableau[2].Nrows- j] << endl; + } + nparticles += RefOffset.Tableau[1].Ncols; + + // cout << "Here c" << endl; + + // Level 3: particles in L part outside GS interval + for (int j = 0; j < RefOffset.Tableau[3].Nrows; ++j) { + Ix2[0][base.Nrap[0] - 1 - RefOffset.Tableau[0].Ncols - RefOffset.Tableau[1].Ncols - RefOffset.Tableau[2].Nrows - j] + = -(base.Nrap[0] + 1 + 2*j + 2*RefOffset.Tableau[3].Row_L[RefOffset.Tableau[3].Nrows - 1 - j]); + //cout << "R pcles, j = " << base.Nrap[0] - 1 - RefOffset.Tableau[0].Ncols - RefOffset.Tableau[1].Ncols - RefOffset.Tableau[2].Nrows << " Ix2[j] = " << Ix2[0][base.Nrap[0] - 1 - RefOffset.Tableau[0].Ncols - RefOffset.Tableau[1].Ncols - RefOffset.Tableau[2].Nrows] << endl; + } + nparticles += RefOffset.Tableau[3].Nrows; + + //cout << "Here d" << endl; + + // Subsequent levels: particles on R and L + for (int base_level = 1; base_level < base.Nrap.size(); ++base_level) { + + for (int j = 0; j < RefOffset.Tableau[2*base_level + 2].Nrows; ++j) + Ix2[base_level][base.Nrap[base_level] - 1 - j] + = ((base.Nrap[base_level] + 1) % 2) + 2*(RefOffset.Tableau[2*base_level + 2].Nrows - 1) -2*j + 2*RefOffset.Tableau[2*base_level + 2].Row_L[j]; + + nparticles += RefOffset.Tableau[2*base_level + 2].Nrows; + + for (int j = 0; j < RefOffset.Tableau[2*base_level + 3].Nrows; ++j) + Ix2[base_level][base.Nrap[base_level] - 1 - RefOffset.Tableau[2*base_level + 2].Nrows - j] + = -1 - ((base.Nrap[base_level]) % 2) -2*j - 2*RefOffset.Tableau[2*base_level + 3].Row_L[RefOffset.Tableau[2*base_level + 3].Nrows - 1 - j]; + + nparticles += RefOffset.Tableau[2*base_level + 3].Nrows; + } + + //// Set the correct parity: quantum numbers always integer. Comparing ODSLF with Heis: if M is even, shift all Ix2 by +1: + //if (base.Nrap[0] % 2 == 0) + //for (int j = 0; j < base.Nrap[0]; ++j) Ix2[0][j] += 1; WRONG!! + + // ODSLF: put back correct parity of quantum nr: + for (int j = 0; j < base.Nrap.size(); ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) + if ((Ix2[j][alpha] + chain.Nsites - 1 + base.Nrap[j] + (base.Nraptot + 1) * chain.Str_L[j] + (chain.par[j] == 1 ? 0 : chain.Nsites)) % 2) + Ix2[j][alpha] -= 1; + //cout << "Here e" << endl; + + //id = RefOffset.id; + //maxid = RefOffset.maxid; + type_id = RefOffset.type_id; + id = RefOffset.id; + maxid = RefOffset.maxid; + + /* + // New meaning of negative parity: + for (int j = 1; j < base.Nrap.size(); ++j) + if (chain.par[j] == -1) // we `invert' the meaning of the quantum numbers + for (int alpha = 0; alpha < base[j]; ++alpha) { + if (Ix2[j][alpha] >= 0) Ix2[j][alpha] -= 2*(base.Ix2_max[j]/2 + 1); + else Ix2[j][alpha] += 2*(base.Ix2_max[j]/2 + 1); + } + */ + + //cout << "After setting Ix2 with offsets of type_id " << type_id << " and id " << id << endl; + //cout << RefOffset.Tableau[0].Nrows << "\t" << RefOffset.Tableau[0].Ncols << "\t" << RefOffset.Tableau[1].Nrows << "\t" << RefOffset.Tableau[1].Ncols << "\t" << RefOffset.Tableau[2].Nrows << "\t" << RefOffset.Tableau[2].Ncols << "\t" << RefOffset.Tableau[3].Nrows << "\t" << RefOffset.Tableau[3].Ncols << "\t" << endl; + //cout << Ix2[0] << endl; + return; + } + + void ODSLF_Bethe_State::Set_to_id (long long int id_ref) + { + //cout << "Heis: Set_to_id called." << endl; + offsets.Set_to_id (id_ref); + (*this).Set_Ix2_Offsets (offsets); + } + + void ODSLF_Bethe_State::Set_to_id (long long int id_ref, ODSLF_Bethe_State& RefState) + { + //cout << "Heis: Set_to_id called." << endl; + offsets.Set_to_id (id_ref); + (*this).Set_Ix2_Offsets (offsets); + } + + bool ODSLF_Bethe_State::Check_Symmetry () + { + // Checks whether the I's are symmetrically distributed. + + bool symmetric_state = true; + int arg, test1, test3; + + //if (chain.Delta <= 1.0) { + + for (int j = 0; j < chain.Nstrings; ++j) { + test1 = 0; + test3 = 0; + for (int alpha = 0; alpha < base[j]; ++alpha) { + arg = (Ix2[j][alpha] != chain.Nsites) ? Ix2[j][alpha] : 0; // since Ix2 = N is same as Ix2 = -N by periodicity, this is symmetric. + test1 += arg; + test3 += arg * arg * arg; // to make sure that all I's are symmetrical... + } + if (!(symmetric_state && (test1 == 0) && (test3 == 0))) symmetric_state = false; + } + //} + /* + else if (chain.Delta > 1.0) { + // For the gapped antiferromagnet, we check symmetry *excluding* the Ix2_max values + // The state is then inadmissible is symmetric && -Ix2_max occupied + for (int j = 0; j < chain.Nstrings; ++j) { + test1 = 0; + test3 = 0; + for (int alpha = 0; alpha < base[j]; ++alpha) { + arg = (Ix2[j][alpha] != chain.Nsites + && abs(Ix2[j][alpha]) != base.Ix2_max[j]) ? Ix2[j][alpha] : 0; // since Ix2 = N is same as Ix2 = -N by periodicity, this is symmetric. + test1 += arg; + test3 += arg * arg * arg; // to make sure that all I's are symmetrical... + } + if (!(symmetric_state && (test1 == 0) && (test3 == 0))) symmetric_state = false; + } + } + */ + + return symmetric_state; + } + + void ODSLF_Bethe_State::Compute_diffsq () + { + DP maxterm = 0.0; + int jmax, alphamax; + + diffsq = 0.0; + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + diffsq += pow(BE[j][alpha], 2.0); + if (pow(BE[j][alpha], 2.0)/chain.Nsites > maxterm) { + jmax = j; + alphamax = alpha; + maxterm = pow(BE[j][alpha], 2.0)/chain.Nsites; + } + } + diffsq /= DP(chain.Nsites); + } + + void ODSLF_Bethe_State::Iterate_BAE () + { + //DP lambda_old; + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) + { + //lambda_old = lambda[j][alpha]; + lambda[j][alpha] = Iterate_BAE (j, alpha); + //cout << j << "\t" << alpha << "\t" << Ix2[j][alpha] << "\t" << lambda_old << "\t" << lambda[j][alpha] << "\t"; + } + //cout << endl; + } + + iter++; + + (*this).Compute_BE(); + (*this).Compute_diffsq(); + } + + void ODSLF_Bethe_State::BAE_smackdown (DP max_allowed) + { + // Re-solves for all rapidities lambda[j][alpha] such that BE[j][alpha]^2/N > max_allowed. + // Assumes that BE[][] is up-to-date. + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) + if (pow(BE[j][alpha], 2.0)/chain.Nsites > max_allowed) (*this).Solve_BAE (j, alpha, max_allowed, 100); + } + + void ODSLF_Bethe_State::Solve_BAE_smackdown (DP max_allowed, int maxruns) + { + int runs_done = 0; + + (*this).Compute_BE(); + (*this).Compute_diffsq(); + + while (diffsq > chain.prec && diffsq > max_allowed && runs_done < maxruns) { + (*this).BAE_smackdown (max_allowed); + (*this).Compute_BE(); + (*this).Compute_diffsq(); + runs_done++; + } + } + + void ODSLF_Bethe_State::Iterate_BAE_Newton () + { + // does one step of a Newton method on the rapidities... + // Assumes that BE[j][alpha] have been computed + + Vect_CX dlambda (0.0, base.Nraptot); // contains delta lambda computed from Newton's method + SQMat_CX Gaudin (0.0, base.Nraptot); + Vect_INT indx (base.Nraptot); + + ODSLF_Lambda lambda_ref(chain, base); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda_ref[j][alpha] = lambda[j][alpha]; + + (*this).Build_Reduced_Gaudin_Matrix (Gaudin); + + int index = 0; + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + dlambda[index] = - BE[j][alpha] * chain.Nsites; + index++; + } + + DP d; + + try { + ludcmp_CX (Gaudin, indx, d); + lubksb_CX (Gaudin, indx, dlambda); + } + + catch (Divide_by_zero) { + diffsq = log(-1.0); // reset to nan to stop Newton method + return; + } + + //cout << iter_Newton << "\t" << dlambda << endl; + + // Regularize dlambda: max step is +-1.0 to prevent rapidity flying off into outer space. + for (int i = 0; i < base.Nraptot; ++i) if (fabs(real(dlambda[i])) > 1.0) dlambda[i] = 0.0;//(real(dlambda[i]) > 0) ? 1.0 : -1.0; + + index = 0; + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + lambda[j][alpha] = lambda_ref[j][alpha] + real(dlambda[index]); + //cout << j << "\t" << alpha << "\t" << dlambda[index] << "\t" << lambda_ref[j][alpha] << "\t" << lambda[j][alpha] << endl; + index++; + } + + (*this).Compute_BE(); + (*this).Iterate_BAE(); + (*this).Compute_diffsq(); + + // if we've converged, calculate the norm here, since the work has been done... + + if (diffsq < chain.prec) { + lnnorm = 0.0; + for (int j = 0; j < base.Nraptot; j++) lnnorm += log(abs(Gaudin[j][j])); + } + + iter_Newton++; + + return; + } + + void ODSLF_Bethe_State::Solve_BAE (int j, int alpha, DP req_prec, int itermax) + { + // Finds the root lambda[j][alpha] to precision req_prec + + DP prec_obtained = 1.0; + int niter_here = 0; + + while (prec_obtained > req_prec && niter_here < itermax) { + lambda[j][alpha] = (*this).Iterate_BAE (j, alpha); + (*this).Compute_BE (j, alpha); + prec_obtained = pow(BE[j][alpha], 2.0)/chain.Nsites; + niter_here++; + } + + } + + void ODSLF_Bethe_State::Find_Rapidities (bool reset_rapidities) + { + // This function finds the rapidities of the eigenstate + + lnnorm = -100.0; // sentinel value, recalculated if Newton method used in the last step of iteration. + + ODSLF_Lambda lambda_ref(chain, base); + + diffsq = 1.0; + + if (reset_rapidities) + (*this).Set_Free_lambdas(); + (*this).Compute_BE(); + + iter = 0; + iter_Newton = 0; + + // Start with conventional iterations + + DP iter_prec = 1.0e-2/DP(chain.Nsites); + + clock_t interp_start; + clock_t interp_stop; + clock_t Newton_start; + clock_t Newton_stop; + int iter_interp_start, iter_interp_stop, iter_Newton_start, iter_Newton_stop; + DP diffsq_interp_start, diffsq_interp_stop, diffsq_Newton_start, diffsq_Newton_stop; + + while (diffsq > chain.prec && iter < 300 && iter_Newton < 20) { + + // If we haven't reset, first try a few Newton steps... + //if (!reset_rapidities && iter_Newton == 0) (*this).Solve_BAE_Newton (chain.prec, 10); + + do { + interp_start = clock(); + iter_interp_start = iter; + diffsq_interp_start = diffsq; + + if (diffsq > iter_prec) (*this).Solve_BAE_interp (iter_prec, 10); + + interp_stop = clock(); + iter_interp_stop = iter; + diffsq_interp_stop = diffsq; + + iter_prec = diffsq * 0.001; + + //cout << "Interp: iter " << iter << "\tdiffsq " << diffsq << endl; + + } while (diffsq > chain.prec && !is_nan(diffsq) && diffsq_interp_stop/diffsq_interp_start < 0.001); + + // If we haven't even reached iter_prec, try normal iterations without extrapolations + + if (diffsq > iter_prec) { + (*this).Solve_BAE_smackdown (0.1 * diffsq, 1); + } + + //cout << "Smackdown: iter " << iter << "\tdiffsq " << diffsq << endl; + + // Now try Newton + + Newton_start = clock(); + iter_Newton_start = iter_Newton; + diffsq_Newton_start = diffsq; + + if (diffsq > iter_prec) (*this).Solve_BAE_Newton (chain.prec, 5); + + Newton_stop = clock(); + iter_Newton_stop = iter_Newton; + diffsq_Newton_stop = diffsq; + + //cout << "Newton: iter " << iter_Newton << "\tdiffsq " << diffsq << endl; + + if (diffsq > iter_prec) (*this).Solve_BAE_smackdown (0.1 * diffsq, 1); + + if (diffsq > iter_prec) (*this).Solve_BAE_straight_iter (0.1 * diffsq, 50); + + //cout << "Straight iter: iter " << iter << "\tdiffsq " << diffsq << endl; + + // If we haven't converged here, retry with more precision in interp method... + + if (is_nan(diffsq)) { + (*this).Set_Free_lambdas(); // good start if we've messed up with Newton + diffsq = 1.0; + } + + iter_prec *= 1.0e-4; + iter_prec = JSC::max(iter_prec, chain.prec); + + } + + // Check convergence: + + //cout << "Check_Rapidities: " << (*this).Check_Rapidities() << endl; + + conv = (diffsq < chain.prec && (*this).Check_Rapidities()) ? 1 : 0; + + return; + } + + void ODSLF_Bethe_State::Solve_BAE_Newton (DP Newton_prec, int max_iter_Newton) + { + // This function attempts to get convergence diffsq <= Newton_prec in at most max_iter_Newton steps. + + // The results are accepted if diffsq has decreased, otherwise the lambda's are reset to original values, defined as... + + ODSLF_Lambda lambda_ref(chain, base); + DP diffsq_ref = 1.0; + + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda_ref[j][alpha] = lambda[j][alpha]; + diffsq_ref = diffsq; + + // Now begin solving... + + int iter_done_here = 0; + + while ((diffsq > Newton_prec) && (diffsq < 10.0) && (!is_nan(diffsq)) && (iter_done_here < max_iter_Newton)) { + (*this).Iterate_BAE_Newton(); + iter_done_here++; + (*this).Iterate_BAE(); + } + + if ((diffsq > diffsq_ref) || (is_nan(diffsq))) { + // This procedure has failed. We reset everything to begin values. + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda[j][alpha] = lambda_ref[j][alpha]; + diffsq = diffsq_ref; + } + + return; + } + + void ODSLF_Bethe_State::Solve_BAE_straight_iter (DP interp_prec, int max_iter_interp) + { + // This function attempts to get convergence diffsq <= interp_prec in at most max_iter_interp steps. + + // If this fails, the lambda's are reset to original values, defined as... + + ODSLF_Lambda lambda_ref(chain, base); + DP diffsq_ref = 1.0; + + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda_ref[j][alpha] = lambda[j][alpha]; + diffsq_ref = diffsq; + + // Now begin solving... + + diffsq = 1.0; + int iter_done_here = 0; + + while ((diffsq > interp_prec) && (max_iter_interp > iter_done_here)) { + (*this).Iterate_BAE(); + iter_done_here++; + } + + if ((diffsq > diffsq_ref)) { + // This procedure has failed. We reset everything to begin values. + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda[j][alpha] = lambda_ref[j][alpha]; + diffsq = diffsq_ref; + } + } + + void ODSLF_Bethe_State::Solve_BAE_interp (DP interp_prec, int max_iter_interp) + { + // This function attempts to get convergence diffsq <= interp_prec in at most max_iter_interp steps. + + // If this fails, the lambda's are reset to original values, defined as... + + ODSLF_Lambda lambda_ref(chain, base); + DP diffsq_ref = 1.0; + + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda_ref[j][alpha] = lambda[j][alpha]; + diffsq_ref = diffsq; + + // Now begin solving... + + diffsq = 1.0; + int iter_done_here = 0; + + ODSLF_Lambda lambda1(chain, base); + ODSLF_Lambda lambda2(chain, base); + ODSLF_Lambda lambda3(chain, base); + ODSLF_Lambda lambda4(chain, base); + ODSLF_Lambda lambda5(chain, base); + + while ((diffsq > interp_prec) && (max_iter_interp > iter_done_here)) { + + (*this).Iterate_BAE(); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda1[j][alpha] = lambda[j][alpha]; + (*this).Iterate_BAE(); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda2[j][alpha] = lambda[j][alpha]; + (*this).Iterate_BAE(); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda3[j][alpha] = lambda[j][alpha]; + (*this).Iterate_BAE(); + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda4[j][alpha] = lambda[j][alpha]; + //(*this).Iterate_BAE(); + //for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda5[j][alpha] = lambda[j][alpha]; + + iter_done_here += 4; + + // now extrapolate the result to infinite number of iterations... + + Vect_DP rap(0.0, 4); + Vect_DP oneoverP(0.0, 4); + DP deltalambda = 0.0; + + for (int i = 0; i < 4; ++i) oneoverP[i] = 1.0/(1.0 + i*i); + + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) { + rap[0] = lambda1[j][alpha]; + rap[1] = lambda2[j][alpha]; + rap[2] = lambda3[j][alpha]; + rap[3] = lambda4[j][alpha]; + //rap[4] = lambda5[j][alpha]; + + polint (oneoverP, rap, 0.0, lambda[j][alpha], deltalambda); + + //cout << j << "\t" << alpha << "\t" << rap << "\t" << lambda[j][alpha] << "\t" << deltalambda << endl; + } + + (*this).Iterate_BAE(); + } + + if ((diffsq >= diffsq_ref)) { + // This procedure has failed. We reset everything to begin values. + for (int j = 0; j < chain.Nstrings; ++j) for (int alpha = 0; alpha < base[j]; ++alpha) lambda[j][alpha] = lambda_ref[j][alpha]; + diffsq = diffsq_ref; + } + + return; + } + + void ODSLF_Bethe_State::Compute_lnnorm () + { + if (true || lnnorm == -100.0) { // else norm already calculated by Newton method + // Actually, compute anyway to increase accuracy ! + + SQMat_CX Gaudin_Red(base.Nraptot); + + (*this).Build_Reduced_Gaudin_Matrix(Gaudin_Red); + /* + cout << endl << "Gaudin matrix: " << endl; + + for (int j = 0; j < Gaudin_Red.size(); ++j) { + for (int k = 0; k < Gaudin_Red.size(); ++k) cout << Gaudin_Red[j][k] << " "; + cout << endl; + } + cout << endl << endl; + */ + complex lnnorm_check = lndet_LU_CX_dstry(Gaudin_Red); + //cout << "Calculated lnnorm = " << lnnorm_check; + + //lnnorm = real(lndet_LU_CX_dstry(Gaudin_Red)); + lnnorm = real(lnnorm_check); + } + + return; + } + + void ODSLF_Bethe_State::Compute_Momentum () + { + int sum_Ix2 = 0; + DP sum_M = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + sum_M += 0.5 * (1.0 + chain.par[j]) * base[j]; + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum_Ix2 += Ix2[j][alpha]; + } + } + + iK = (chain.Nsites/2) * int(sum_M + 0.1) - (sum_Ix2/2); // + 0.1: for safety... + + while (iK >= chain.Nsites) iK -= chain.Nsites; + while (iK < 0) iK += chain.Nsites; + + K = PI * sum_M - PI * sum_Ix2/chain.Nsites; + + while (K >= 2.0*PI) K -= 2.0*PI; + while (K < 0.0) K += 2.0*PI; + + return; + } + + void ODSLF_Bethe_State::Compute_All (bool reset_rapidities) // solves BAE, computes E, K and lnnorm + { + (*this).Find_Rapidities (reset_rapidities); + if (conv == 1) { + (*this).Compute_Energy (); + (*this).Compute_Momentum (); + (*this).Compute_lnnorm (); + } + return; + } + + bool ODSLF_Bethe_State::Boost_Momentum (int iKboost) + { + if (iKboost == 0) return(true); // done + + ODSLF_Ix2_Offsets offsets_here = offsets; + + bool success = false; + + if (iKboost < 0) + success = offsets_here.Add_Boxes_From_Lowest(-iKboost, 0); // add boxes in even sectors to decrease iK + + else if (iKboost > 0) + success = offsets_here.Add_Boxes_From_Lowest(iKboost, 1); // add boxes in odd sectors to increase iK + + if (success) (*this).Set_Ix2_Offsets(offsets_here); + + return(success); + } + + + std::ostream& operator<< (std::ostream& s, const ODSLF_Bethe_State& state) + { + // sends all the state data to output stream + + s << endl << "******** Chain with Nsites = " << state.chain.Nsites << " Mdown (nr fermions) = " << state.base.Mdown + << ": eigenstate with base_id " << state.base_id << ", type_id " << state.type_id + << " id " << state.id << " maxid " << state.offsets.maxid << endl + << "E = " << state.E << " K = " << state.K << " iK = " << state.iK << " lnnorm = " << state.lnnorm << endl + << "conv = " << state.conv << " iter = " << state.iter << " iter_Newton = " << state.iter_Newton << "\tdiffsq " << state.diffsq << endl; + + for (int j = 0; j < state.chain.Nstrings; ++j) { + if (state.base.Nrap[j] > 0) { + s << "Type " << j << " Str_L = " << state.chain.Str_L[j] << " par = " << state.chain.par[j] << " M_j = " << state.base.Nrap[j] + << " Ix2_infty = " << state.base.Ix2_infty[j] << " Ix2_max = " << state.base.Ix2_max[j] << endl; + Vect_INT qnumbers(state.base.Nrap[j]); + Vect_DP rapidities(state.base.Nrap[j]); + for (int alpha = 0; alpha < state.base.Nrap[j]; ++alpha) { + qnumbers[alpha] = state.Ix2[j][alpha]; + rapidities[alpha] = state.lambda[j][alpha]; + } + qnumbers.QuickSort(); + rapidities.QuickSort(); + + s << "Ix2 quantum numbers: " << endl; + for (int alpha = 0; alpha < state.base.Nrap[j]; ++alpha) s << qnumbers[alpha] << " "; + s << endl; + s << "Rapidities: " << endl; + for (int alpha = 0; alpha < state.base.Nrap[j]; ++alpha) s << rapidities[alpha] << " "; + s << endl; + } + } + s << endl; + + return s; + + } + + + +} // namespace JSC diff --git a/src/ODSLF/ODSLF_Chem_Pot.cc b/src/ODSLF/ODSLF_Chem_Pot.cc new file mode 100644 index 0000000..4c304ae --- /dev/null +++ b/src/ODSLF/ODSLF_Chem_Pot.cc @@ -0,0 +1,144 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/ODSLF/ODSLF_Chem_Pot.cc + +Purpose: calculates the chemical potential. + + +***********************************************************/ + +#include "JSC.h" + +namespace JSC { + + /* + DP Ezero (DP Delta, int N, int M) + { + // Returns the energy of the ground state with M down spins + + if (M < 0 || M > N/2) JSCerror("M out of bounds in Ezero."); + + DP E = -1.0; // sentinel value + + if (M == 0) E = N * Delta/4.0; + + else { + + Heis_Chain BD1(1.0, Delta, 0.0, N); + + Vect_INT Nrapidities_groundstate(0, BD1.Nstrings); + + Nrapidities_groundstate[0] = M; + + ODSLF_Base baseconfig_groundstate(BD1, Nrapidities_groundstate); + + if ((Delta > 0.0) && (Delta < 1.0)) { + ODSLF_XXZ_Bethe_State groundstate(BD1, baseconfig_groundstate); + groundstate.Compute_All(true); + E = groundstate.E; + } + + else if (Delta == 1.0) { + XXX_Bethe_State groundstate(BD1, baseconfig_groundstate); + groundstate.Compute_All(true); + E = groundstate.E; + } + + else if (Delta > 1.0) { + XXZ_gpd_Bethe_State groundstate(BD1, baseconfig_groundstate); + groundstate.Compute_All(true); + E = groundstate.E; + } + + else JSCerror("Anisotropy out of bounds in Ezero."); + } + + return(E); + } + + DP H_vs_M (DP Delta, int N, int M) + { + // Assumes dE/dM = 0 = dE_0/dM + h, with dE_0/dM = E_0(M) - E_0 (M - 1) + + DP H = 0.0; + + if (2*M == N) H = 0.0; + + else if (Delta <= 1.0) H = Ezero (Delta, N, M - 1) - Ezero (Delta, N, M); + + return(H); + } + + DP HZmin (DP Delta, int N, int M, Vect_DP& Ezero_ref) + { + if (M < 0 || M > N/2 - 1) { + cout << "M = " << M << endl; + JSCerror("M out of bounds in HZmin."); + } + + if (Ezero_ref[M] == -1.0) Ezero_ref[M] = Ezero(Delta, N, M); + if (Ezero_ref[M + 1] == -1.0) Ezero_ref[M + 1] = Ezero(Delta, N, M + 1); + + return(Ezero_ref[M] - Ezero_ref[M + 1]); + } + + int M_vs_H (DP Delta, int N, DP HZ) + { + // Returns the value of M for given field HZ + + if (HZ < 0.0) JSCerror("Please use a positive field in M_vs_H."); + + else if (HZ == 0.0) return(N/2); + + // Here, -1.0 is a sentinel value. + Vect_DP Ezero(-1.0, N/2 + 1); // contains the GSE[M]. + + // We look for M s.t. HZmin[M] < HZ <= HZmin[M + 1] + + int M_actual = N/4; // start somewhere in middle + int M_step = N/8 - 1; // step + DP HZmin_actual = 0.0; + DP HZmax_actual = 0.0; + bool M_found = false; + + if (HZ >= 1.0 + Delta) M_actual = 0; // saturation + + else { + + HZmin_actual = HZmin (Delta, N, M_actual, Ezero); + HZmax_actual = HZmin (Delta, N, M_actual - 1, Ezero); + + while (!M_found) { + + if (HZmin_actual > HZ) M_actual += M_step; + else if (HZmax_actual <= HZ) M_actual -= M_step; + + M_step = (M_step + 1)/2; + + HZmin_actual = HZmin (Delta, N, M_actual, Ezero); + HZmax_actual = HZmin (Delta, N, M_actual - 1, Ezero); + + M_found = (HZmin_actual < HZ && HZ <= HZmax_actual); + + //cout << "M_actual = " << M_actual << "\tM_step = " << M_step + // << "\tHZmin_actual = " << HZmin_actual << "\tHZmax_actual = " << HZmax_actual << "\tHZ = " << HZ << "\t" << M_found << endl; + } + } + //cout << "M found = " << M_actual << "\tHZmax = " << Ezero[M_actual] - Ezero[M_actual + 1] << "\tHZmin = " << Ezero[M_actual - 1] - Ezero[M_actual] << endl; + + return(M_actual); + } + */ + + DP Chemical_Potential (const ODSLF_Bethe_State& RefState) + { + return(-H_vs_M (RefState.chain.Delta, RefState.chain.Nsites, RefState.base.Mdown)); // - sign since E_{M+1} - E_M = -H + } + +} diff --git a/src/ODSLF/ODSLF_Matrix_Element_Contrib.cc b/src/ODSLF/ODSLF_Matrix_Element_Contrib.cc new file mode 100644 index 0000000..29afa3c --- /dev/null +++ b/src/ODSLF/ODSLF_Matrix_Element_Contrib.cc @@ -0,0 +1,87 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/ODSLF/ODSLF_Matrix_Element_Contrib.cc + +Purpose: handles the generic call for a matrix element. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + //DP Compute_Matrix_Element_Contrib (char whichDSF, bool fixed_iK, ODSLF_XXZ_Bethe_State& LeftState, + //ODSLF_XXZ_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile) + DP Compute_Matrix_Element_Contrib (char whichDSF, int iKmin, int iKmax, ODSLF_XXZ_Bethe_State& LeftState, + ODSLF_XXZ_Bethe_State& RefState, DP Chem_Pot, fstream& DAT_outfile) + { + // This function prints the matrix element information to the fstream, + // and returns a weighed `data_value' to be multiplied by sumrule_factor, + // to determine if scanning along this thread should be continued. + + // Identify which matrix is needed from the number of particles in Left and Right states: + + bool fixed_iK = (iKmin == iKmax); + + DP ME = 0.0; + if (whichDSF == 'Z') + ME = LeftState.E - RefState.E; + else if (whichDSF == 'm') + ME = exp(real(ln_Smin_ME (RefState, LeftState))); + else if (whichDSF == 'z') { + if (LeftState.base_id == RefState.base_id && LeftState.type_id == RefState.type_id && LeftState.id == RefState.id) + //MEsq = RefState.chain.Nsites * 0.25 * pow((1.0 - 2.0*RefState.base.Mdown/RefState.chain.Nsites), 2.0); + ME = sqrt(RefState.chain.Nsites * 0.25) * (1.0 - 2.0*RefState.base.Mdown/RefState.chain.Nsites); + else ME = exp(real(ln_Sz_ME (RefState, LeftState))); + } + else if (whichDSF == 'p') + ME = exp(real(ln_Smin_ME (LeftState, RefState))); + else JSCerror("Wrong whichDSF in Compute_Matrix_Element_Contrib."); + + if (is_nan(ME)) ME = 0.0; + + // Do the momentum business: + int iKout = LeftState.iK - RefState.iK; + while(iKout < 0) iKout += RefState.chain.Nsites; + while(iKout >= RefState.chain.Nsites) iKout -= RefState.chain.Nsites; + + // Print information to fstream: + if (iKout >= iKmin && iKout <= iKmax) { + if (whichDSF == 'Z') { + DAT_outfile << endl << LeftState.E - RefState.E - (LeftState.base.Mdown - RefState.base.Mdown) * Chem_Pot << "\t" + << iKout << "\t" + //<< LeftState.conv << "\t" + << LeftState.base_id << "\t" << LeftState.type_id << "\t" << LeftState.id; + } + else { + DAT_outfile << endl << LeftState.E - RefState.E - (LeftState.base.Mdown - RefState.base.Mdown) * Chem_Pot << "\t" + << iKout << "\t" + << ME << "\t" + //<< LeftState.conv << "\t" + << LeftState.base_id << "\t" << LeftState.type_id << "\t" << LeftState.id; + } + } // if iKmin <= iKout <= iKmax + + // Calculate and return the data_value: + DP data_value = ME * ME; + //DP data_value = (iKout == 0 ? 1.0 : 2.0) * MEsq; + if (whichDSF == 'Z') // use 1/(1 + omega) + data_value = 1.0/(1.0 + LeftState.E - RefState.E - (LeftState.base.Mdown - RefState.base.Mdown) * Chem_Pot); + else if (fixed_iK) // data value is MEsq * omega: + data_value = ME * ME * (LeftState.E - RefState.E - (LeftState.base.Mdown - RefState.base.Mdown) * Chem_Pot); + + return(data_value); + } + + +} // namespace JSC diff --git a/src/ODSLF/ODSLF_Sumrules.cc b/src/ODSLF/ODSLF_Sumrules.cc new file mode 100644 index 0000000..aee87a2 --- /dev/null +++ b/src/ODSLF/ODSLF_Sumrules.cc @@ -0,0 +1,245 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/ODSLF/ODSLF_Sumrules.cc + +Purpose: defines sumrule factors for spinless fermions related to Heisenberg + + +***********************************************************/ + +#include "JSC.h" + +using namespace JSC; +using namespace std; + +namespace JSC { + + + DP ODSLF_X_avg (char xyorz, DP Delta, int N, int M) + { + // Calculates \sum_j < S_j^a S_{j+1}^a >, a = x, y or z. + + DP eps_Delta = 0.00000001; + + // Define the chain: J, Delta, h, Nsites + Heis_Chain chain(1.0, Delta, 0.0, N); + + // Define the base: chain, Mdown + ODSLF_Base gbase(chain, M); + + // Define the chain: J, Delta, h, Nsites + Heis_Chain chain2(1.0, Delta + eps_Delta, 0.0, N); + + // Define the base: chain, Mdown + ODSLF_Base gbase2(chain2, M); + + DP E0_Delta = 0.0; + DP E0_Delta_eps = 0.0; + + if (Delta > 0.0 && Delta < 1.0) { + + // Define the ground state + ODSLF_XXZ_Bethe_State gstate(chain, gbase); + + // Compute everything about the ground state + gstate.Compute_All(true); + + E0_Delta = gstate.E; + + // Define the ground state + ODSLF_XXZ_Bethe_State gstate2(chain2, gbase2); + + // Compute everything about the ground state + gstate2.Compute_All(true); + + E0_Delta_eps = gstate2.E; + } + /* + else if (Delta == 1.0) { + // Define the ground state + XXX_Bethe_State gstate(chain, gbase); + + // Compute everything about the ground state + gstate.Compute_All(true); + + E0_Delta = gstate.E; + + // Define the ground state + XXZ_gpd_Bethe_State gstate2(chain2, gbase2); // need XXZ_gpd here + + // Compute everything about the ground state + gstate2.Compute_All(true); + + E0_Delta_eps = gstate2.E; + } + + else if (Delta > 1.0) { + // Define the ground state + XXZ_gpd_Bethe_State gstate(chain, gbase); + + // Compute everything about the ground state + gstate.Compute_All(true); + + E0_Delta = gstate.E; + + // Define the ground state + XXZ_gpd_Bethe_State gstate2(chain2, gbase2); + + // Compute everything about the ground state + gstate2.Compute_All(true); + + E0_Delta_eps = gstate2.E; + } + */ + else JSCerror("Wrong anisotropy in ODSLF_S1_sumrule_factor."); + + DP answer = 0.0; + if (xyorz == 'x' || xyorz == 'y') answer = 0.5 * (E0_Delta - Delta * (E0_Delta_eps - E0_Delta)/eps_Delta); + + // Careful for z ! Hamiltonian defined as S^z S^z - 1/4, so add back N/4: + else if (xyorz == 'z') answer = (E0_Delta_eps - E0_Delta)/eps_Delta + 0.25 * N; + + else JSCerror("option not implemented in ODSLF_X_avg."); + + return(answer); + } + + DP ODSLF_S1_sumrule_factor (char mporz, DP Delta, int N, int M, int iK) + { + + DP X_x = ODSLF_X_avg ('x', Delta, N, M); + DP X_z = ODSLF_X_avg ('z', Delta, N, M); + + DP sumrule = 0.0; + + if (mporz == 'm' || mporz == 'p') sumrule = - 2.0 * ((1.0 - Delta * cos((twoPI * iK)/N)) * X_x + (Delta - cos((twoPI * iK)/N)) * X_z)/N; + //if (mporz == 'm' || mporz == 'p') sumrule = - 1.0 * ((1.0 - Delta * cos((twoPI * iK)/N)) * 2.0 * X_x + (Delta - cos((twoPI * iK)/N)) * (X_x + X_z))/N; + + else if (mporz == 'z') sumrule = iK == 0 ? 1.0 : -2.0 * X_x * (1.0 - cos((twoPI * iK)/N))/N; + + else if (mporz == 'a') sumrule = 1.0; + else if (mporz == 'b') sumrule = 1.0; + + else JSCerror("option not implemented in ODSLF_S1_sumrule_factor."); + + //return(1.0/sumrule); + return(1.0/(sumrule + 1.0e-16)); // sumrule is 0 for iK == 0 or N + } + + DP ODSLF_S1_sumrule_factor (char mporz, DP Delta, int N, DP X_x, DP X_z, int iK) + { + + //DP X_x = X_avg ('x', Delta, N, M); + //DP X_z = X_avg ('z', Delta, N, M); + + DP sumrule = 0.0; + + if (mporz == 'm' || mporz == 'p') sumrule = - 2.0 * ((1.0 - Delta * cos((twoPI * iK)/N)) * X_x + (Delta - cos((twoPI * iK)/N)) * X_z)/N; + //if (mporz == 'm' || mporz == 'p') sumrule = - 1.0 * ((1.0 - Delta * cos((twoPI * iK)/N)) * 2.0 * X_x + (Delta - cos((twoPI * iK)/N)) * (X_x + X_z))/N; + + else if (mporz == 'z') sumrule = -2.0 * X_x * (1.0 - cos((twoPI * iK)/N))/N; + + else if (mporz == 'a') sumrule = 1.0; + else if (mporz == 'b') sumrule = 1.0; + + else JSCerror("option not implemented in ODSLF_S1_sumrule_factor."); + + //return(1.0/sumrule); + return(1.0/(sumrule + 1.0e-16)); // sumrule is 0 for iK == 0 or N + } + + //DP Sumrule_Factor (char whichDSF, Heis_Bethe_State& RefState, DP Chem_Pot, bool fixed_iK, int iKneeded) + DP Sumrule_Factor (char whichDSF, ODSLF_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax) + { + DP sumrule_factor = 1.0; + //if (!fixed_iK) { + if (iKmin != iKmax) { + if (whichDSF == 'Z') sumrule_factor = 1.0; + else if (whichDSF == 'm') + sumrule_factor = 1.0/RefState.base.Mdown; + else if (whichDSF == 'z') sumrule_factor = 1.0/(0.25 * RefState.chain.Nsites); + else if (whichDSF == 'p') sumrule_factor = 1.0/(RefState.chain.Nsites - RefState.base.Mdown); + else if (whichDSF == 'a') sumrule_factor = 1.0; + else if (whichDSF == 'b') sumrule_factor = 1.0; + else if (whichDSF == 'q') sumrule_factor = 1.0; + + else JSCerror("whichDSF option not consistent in Sumrule_Factor"); + } + //else if (fixed_iK) { + else if (iKmin == iKmax) { + if (whichDSF == 'Z') sumrule_factor = 1.0; + else if (whichDSF == 'm' || whichDSF == 'z' || whichDSF == 'p') + //sumrule_factor = S1_sumrule_factor (whichDSF, RefState.chain.Delta, RefState.chain.Nsites, RefState.base.Mdown, iKneeded); + sumrule_factor = ODSLF_S1_sumrule_factor (whichDSF, RefState.chain.Delta, RefState.chain.Nsites, RefState.base.Mdown, iKmax); + else if (whichDSF == 'a') sumrule_factor = 1.0; + else if (whichDSF == 'b') sumrule_factor = 1.0; + else if (whichDSF == 'q') sumrule_factor = 1.0; + + else JSCerror("whichDSF option not consistent in Sumrule_Factor"); + } + + + + return(sumrule_factor); + } + + //void Evaluate_F_Sumrule (char whichDSF, const ODSLF_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax, const char* FFsq_Cstr, const char* FSR_Cstr) + void Evaluate_F_Sumrule (string prefix, char whichDSF, const ODSLF_Bethe_State& RefState, DP Chem_Pot, int iKmin, int iKmax) + { + + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + stringstream FSR_stringstream; string FSR_string; + FSR_stringstream << prefix << ".fsr"; + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + + ifstream infile; + infile.open(RAW_Cstr); + if(infile.fail()) JSCerror("Could not open input file in Evaluate_F_Sumrule(ODSLF...)."); + + // We run through the data file to chech the f sumrule at each positive momenta: + //Vect Sum_omega_FFsq(0.0, RefState.chain.Nsites/2 + 1); // + Vect Sum_omega_FFsq(0.0, iKmax - iKmin + 1); // + + DP omega, FF; + int iK, conv; + long long int base_id, type_id, id; + + while (infile.peek() != EOF) { + infile >> omega >> iK >> FF >> conv >> base_id >> type_id >> id; + //if (iK > 0 && iK <= RefState.chain.Nsites/2) Sum_omega_FFsq[iK] += omega * FFsq; + if (iK >= iKmin && iK <= iKmax) Sum_omega_FFsq[iK - iKmin] += omega * FF * FF; + } + + infile.close(); + + ofstream outfile; + outfile.open(FSR_Cstr); + outfile.precision(16); + + DP X_x = X_avg ('x', RefState.chain.Delta, RefState.chain.Nsites, RefState.base.Mdown); + DP X_z = X_avg ('z', RefState.chain.Delta, RefState.chain.Nsites, RefState.base.Mdown); + + /* + outfile << 0 << "\t" << Sum_omega_FFsq[0] * S1_sumrule_factor (whichDSF, RefState.chain.Delta, RefState.chain.Nsites, X_x, X_z, 0); + for (int i = 1; i <= RefState.chain.Nsites/2; ++i) + outfile << endl << i << "\t" << Sum_omega_FFsq[i] * S1_sumrule_factor (whichDSF, RefState.chain.Delta, RefState.chain.Nsites, X_x, X_z, i); + */ + + for (int i = iKmin; i <= iKmax; ++i) { + if (i > iKmin) outfile << endl; + outfile << i << "\t" << Sum_omega_FFsq[i] * ODSLF_S1_sumrule_factor (whichDSF, RefState.chain.Delta, RefState.chain.Nsites, X_x, X_z, i); + } + + outfile.close(); + } + +} // namespace JSC diff --git a/src/ODSLF/ODSLF_XXZ_Bethe_State.cc b/src/ODSLF/ODSLF_XXZ_Bethe_State.cc new file mode 100644 index 0000000..009ec79 --- /dev/null +++ b/src/ODSLF/ODSLF_XXZ_Bethe_State.cc @@ -0,0 +1,496 @@ +/**************************************************************** + +ODSLF_XXZ_Bethe_State.cc + +Defines all functions for ODSLF_XXZ_Bethe_State + + +******************************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + // Function prototypes + + inline DP ODSLF_fbar_XXZ (DP lambda, int par, DP tannzetaover2); + DP ODSLF_Theta_XXZ (DP lambda, int nj, int nk, int parj, int park, DP* tannzetaover2); + DP ODSLF_hbar_XXZ (DP lambda, int n, int par, DP* si_n_anis_over_2); + DP ODSLF_ddlambda_Theta_XXZ (DP lambda, int nj, int nk, int parj, int park, DP* si_n_anis_over_2); + + + //*************************************************************************************************** + + // Function definitions: class ODSLF_XXZ_Bethe_State + + ODSLF_XXZ_Bethe_State::ODSLF_XXZ_Bethe_State () + : ODSLF_Bethe_State(), sinhlambda(ODSLF_Lambda(chain, 1)), coshlambda(ODSLF_Lambda(chain, 1)), tanhlambda(ODSLF_Lambda(chain, 1)) + {}; + + ODSLF_XXZ_Bethe_State::ODSLF_XXZ_Bethe_State (const ODSLF_XXZ_Bethe_State& RefState) // copy constructor + : ODSLF_Bethe_State(RefState), sinhlambda(ODSLF_Lambda(RefState.chain, RefState.base)), coshlambda(ODSLF_Lambda(RefState.chain, RefState.base)), + tanhlambda(ODSLF_Lambda(RefState.chain, RefState.base)) + { + // copy arrays into new ones + + //cout << "Calling XXZ state copy constructor." << endl; + for (int j = 0; j < RefState.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < RefState.base[j]; ++j) { + sinhlambda[j][alpha] = RefState.sinhlambda[j][alpha]; + coshlambda[j][alpha] = RefState.coshlambda[j][alpha]; + tanhlambda[j][alpha] = RefState.tanhlambda[j][alpha]; + } + } + //cout << "Done calling XXZ state copy constructor." << endl; + } + + ODSLF_XXZ_Bethe_State::ODSLF_XXZ_Bethe_State (const Heis_Chain& RefChain, int M) + : ODSLF_Bethe_State(RefChain, M), + sinhlambda(ODSLF_Lambda(RefChain, M)), coshlambda(ODSLF_Lambda(RefChain, M)), tanhlambda(ODSLF_Lambda(RefChain, M)) + { + //cout << "Here in XXZ BS constructor." << endl; + //cout << (*this).lambda[0][0] << endl; + //cout << "OK" << endl; + if ((RefChain.Delta <= -1.0) || (RefChain.Delta >= 1.0)) JSCerror("Delta out of range in ODSLF_XXZ_Bethe_State constructor"); + } + + ODSLF_XXZ_Bethe_State::ODSLF_XXZ_Bethe_State (const Heis_Chain& RefChain, const ODSLF_Base& RefBase) + : ODSLF_Bethe_State(RefChain, RefBase), + sinhlambda(ODSLF_Lambda(RefChain, RefBase)), coshlambda(ODSLF_Lambda(RefChain, RefBase)), tanhlambda(ODSLF_Lambda(RefChain, RefBase)) + { + if ((RefChain.Delta <= -1.0) || (RefChain.Delta >= 1.0)) JSCerror("Delta out of range in ODSLF_XXZ_Bethe_State constructor"); + } + + ODSLF_XXZ_Bethe_State::ODSLF_XXZ_Bethe_State (const Heis_Chain& RefChain, long long int base_id_ref, long long int type_id_ref) + : ODSLF_Bethe_State(RefChain, base_id_ref, type_id_ref), + sinhlambda(ODSLF_Lambda(chain, base)), coshlambda(ODSLF_Lambda(chain, base)), tanhlambda(ODSLF_Lambda(chain, base)) + { + if ((RefChain.Delta <= -1.0) || (RefChain.Delta >= 1.0)) JSCerror("Delta out of range in ODSLF_XXZ_Bethe_State constructor"); + } + + + ODSLF_XXZ_Bethe_State& ODSLF_XXZ_Bethe_State::operator= (const ODSLF_XXZ_Bethe_State& RefState) + { + if (this != &RefState) { + chain = RefState.chain; + base = RefState.base; + offsets = RefState.offsets; + Ix2 = RefState.Ix2; + lambda = RefState.lambda; + BE = RefState.BE; + diffsq = RefState.diffsq; + conv = RefState.conv; + iter = RefState.iter; + iter_Newton = RefState.iter_Newton; + E = RefState.E; + iK = RefState.iK; + K = RefState.K; + lnnorm = RefState.lnnorm; + base_id = RefState.base_id; + type_id = RefState.type_id; + id = RefState.id; + maxid = RefState.maxid; + nparticles = RefState.nparticles; + + sinhlambda = RefState.sinhlambda; + coshlambda = RefState.coshlambda; + tanhlambda = RefState.tanhlambda; + } + return(*this); + } + + // Member functions + + void ODSLF_XXZ_Bethe_State::Set_Free_lambdas() + { + // Sets all the rapidities to the solutions of the BAEs without scattering terms + + for (int i = 0; i < chain.Nstrings; ++i) { + + for (int alpha = 0; alpha < base[i]; ++alpha) { + + if(chain.par[i] == 1) lambda[i][alpha] = (tan(chain.Str_L[i] * 0.5 * chain.anis) * tan(PI * 0.5 * Ix2[i][alpha]/chain.Nsites)); + else if (chain.par[i] == -1) lambda[i][alpha] = (-tan((PI * 0.5 * Ix2[i][alpha])/chain.Nsites)/tan(chain.Str_L[i] * 0.5 * chain.anis)); + + else JSCerror("Invalid parities in Set_Free_lambdas."); + + // if (lambda[i][alpha] == 0.0) lambda[i][alpha] = 0.001 * (1.0 + i) * (1.0 + alpha) / chain.Nsites; // some arbitrary starting point here... + + } + } + + return; + } + + void ODSLF_XXZ_Bethe_State::Compute_sinhlambda() + { + for (int j = 0; j < chain.Nstrings; ++j) { + + for (int alpha = 0; alpha < base[j]; ++alpha) sinhlambda[j][alpha] = sinh(lambda[j][alpha]); + } + return; + } + + void ODSLF_XXZ_Bethe_State::Compute_coshlambda() + { + for (int j = 0; j < chain.Nstrings; ++j) { + + for (int alpha = 0; alpha < base[j]; ++alpha) coshlambda[j][alpha] = cosh(lambda[j][alpha]); + } + return; + } + + void ODSLF_XXZ_Bethe_State::Compute_tanhlambda() + { + for (int j = 0; j < chain.Nstrings; ++j) { + + for (int alpha = 0; alpha < base[j]; ++alpha) tanhlambda[j][alpha] = tanh(lambda[j][alpha]); + } + return; + } + + bool ODSLF_XXZ_Bethe_State::Check_Admissibility(char option) + { + // This function checks the admissibility of the Ix2's of a state: + // returns false if there are higher strings with Ix2 = 0, a totally symmetric distribution of I's at each level, + // and strings of equal length modulo 2 and parity with Ix2 = 0, meaning at least two equal roots in BAE. + + bool answer = true; + Vect Zero_at_level(false, chain.Nstrings); // whether there exists an Ix2 == 0 at a given level + + bool higher_string_on_zero = false; + for (int j = 0; j < chain.Nstrings; ++j) { + // The following line puts answer to true if there is at least one higher string with zero Ix2 + for (int alpha = 0; alpha < base[j]; ++alpha) if ((Ix2[j][alpha] == 0) && (chain.Str_L[j] >= 2) /*&& !(chain.Str_L[j] % 2)*/) + higher_string_on_zero = true; + for (int alpha = 0; alpha < base[j]; ++alpha) if (Ix2[j][alpha] == 0) Zero_at_level[j] = true; + // NOTE: if base[j] == 0, Zero_at_level[j] remains false. + } + + // check symmetry of Ix2 at each level, if there exists a potentially risky Ix2... + + bool symmetric_state = (*this).Check_Symmetry(); + + bool string_coincidence = false; + for (int j1 = 0; j1 < chain.Nstrings; ++j1) { + for (int j2 = j1 + 1; j2 < chain.Nstrings; ++j2) + if (Zero_at_level[j1] && Zero_at_level[j2] && (chain.par[j1] == chain.par[j2]) && (!((chain.Str_L[j1] + chain.Str_L[j2])%2))) + string_coincidence = true; + } + + bool M_odd_and_onep_on_zero = false; + if (option == 'z') { // for Sz, if M is odd, exclude symmetric states with a 1+ on zero + // (zero rapidities in left and right states, so FF det not defined). + bool is_ground_state = base.Nrap[0] == base.Mdown && Ix2[0][0] == -(base.Mdown - 1) && Ix2[0][base.Mdown-1] == base.Mdown - 1; + if (Zero_at_level[0] && (base.Mdown % 2) && !is_ground_state) M_odd_and_onep_on_zero = true; + } + + bool onep_onem_on_zero = false; + if (option == 'm' || option == 'p') { // for Smin, we also exclude symmetric states with 1+ and 1- strings on zero + if (Zero_at_level[0] && Zero_at_level[1]) onep_onem_on_zero = true; + } + + answer = !(symmetric_state && (higher_string_on_zero || string_coincidence || onep_onem_on_zero || M_odd_and_onep_on_zero)); + + // Now check that no Ix2 is equal to +N (since we take -N into account, and I + N == I by periodicity of exp) + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) if ((Ix2[j][alpha] < -chain.Nsites) || (Ix2[j][alpha] >= chain.Nsites)) answer = false; + + if (!answer) { + E = 0.0; + K = 0.0; + conv = 0; + iter = 0; + iter_Newton = 0; + lnnorm = -100.0; + } + + return(answer); // answer == true: nothing wrong with this Ix2_config + } + + void ODSLF_XXZ_Bethe_State::Compute_BE (int j, int alpha) + { + tanhlambda[j][alpha] = tanh(lambda[j][alpha]); + + DP sumtheta = 0.0; + + for (int k = 0; k < chain.Nstrings; ++k) + for (int beta = 0; beta < base[k]; ++beta) { + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += (chain.par[j] == chain.par[k]) + ? atan((tanhlambda[j][alpha] - tanhlambda[k][beta])/((1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta]) * chain.ta_n_anis_over_2[2])) + : - atan(((tanhlambda[j][alpha] - tanhlambda[k][beta])/(1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta])) * chain.ta_n_anis_over_2[2]) ; + else sumtheta += 0.5 * ODSLF_Theta_XXZ((tanhlambda[j][alpha] - tanhlambda[k][beta])/(1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.par[j], chain.par[k], chain.ta_n_anis_over_2); + } + sumtheta *= 2.0; + + BE[j][alpha] = ((chain.par[j] == 1) ? 2.0 * atan(tanhlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) + : -2.0 * atan(tanhlambda[j][alpha] * chain.ta_n_anis_over_2[chain.Str_L[j]])) - (sumtheta + PI*Ix2[j][alpha])/chain.Nsites; + + } + + void ODSLF_XXZ_Bethe_State::Compute_BE () + { + // Fills in the BE members with the value of the Bethe equations. + + (*this).Compute_tanhlambda(); + + DP sumtheta = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) { + + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) + for (int beta = 0; beta < base[k]; ++beta) { + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += (chain.par[j] == chain.par[k]) + ? atan((tanhlambda[j][alpha] - tanhlambda[k][beta])/((1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta]) * chain.ta_n_anis_over_2[2])) + : - atan(((tanhlambda[j][alpha] - tanhlambda[k][beta])/(1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta])) * chain.ta_n_anis_over_2[2]) ; + else sumtheta += 0.5 * ODSLF_Theta_XXZ((tanhlambda[j][alpha] - tanhlambda[k][beta])/(1.0 - tanhlambda[j][alpha] * tanhlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.par[j], chain.par[k], chain.ta_n_anis_over_2); + } + sumtheta *= 2.0; + + BE[j][alpha] = ((chain.par[j] == 1) ? 2.0 * atan(tanhlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) + : -2.0 * atan(tanhlambda[j][alpha] * chain.ta_n_anis_over_2[chain.Str_L[j]])) - (sumtheta + PI*Ix2[j][alpha])/chain.Nsites; + + //if (is_nan(BE[j][alpha])) cout << "BE nan: " << j << "\t" << alpha << "\t" << lambda[j][alpha] << "\t" << tanhlambda[j][alpha] << endl; + } + } + + DP ODSLF_XXZ_Bethe_State::Iterate_BAE (int j, int alpha) + { + // Returns a new iteration value for lambda[j][alpha] given tanhlambda and BE Lambdas + // Assumes that tanhlambda[][] and BE[][] have been computed. + + DP new_lambda = 0.0; + DP arg = 0.0; + + if (chain.par[j] == 1) arg = chain.ta_n_anis_over_2[chain.Str_L[j]] + * tan(0.5 * + //(PI * Ix2[j][alpha] + sumtheta)/chain.Nsites + (2.0 * atan(tanhlambda[j][alpha]/chain.ta_n_anis_over_2[chain.Str_L[j]]) - BE[j][alpha]) + ); + + else if (chain.par[j] == -1) arg = -tan(0.5 * + //(PI * Ix2[j][alpha] + sumtheta)/chain.Nsites) + (-2.0 * atan(tanhlambda[j][alpha] * chain.ta_n_anis_over_2[chain.Str_L[j]]) - BE[j][alpha])) + /chain.ta_n_anis_over_2[chain.Str_L[j]]; + + if (fabs(arg) < 1.0) { + new_lambda = atanh(arg); + } + + else { + new_lambda = lambda[j][alpha]; // back to drawing board... + int block = 0; // counter to prevent runaway while loop + DP new_tanhlambda = 0.0; + DP sumtheta = 0.0; + arg = 10.0; // reset value to start while loop + while ((fabs(arg) > 1.0) && (block++ < 100)) { // recompute the diverging root on its own... + new_lambda *= 1.01; // try to go slowly towards infinity... + new_tanhlambda = tanh(new_lambda); + sumtheta = 0.0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + sumtheta += (chain.par[j] == chain.par[k]) + ? atan((new_tanhlambda - tanhlambda[k][beta])/((1.0 - new_tanhlambda * tanhlambda[k][beta]) * chain.ta_n_anis_over_2[2])) + : - atan(((new_tanhlambda - tanhlambda[k][beta])/(1.0 - new_tanhlambda * tanhlambda[k][beta])) * chain.ta_n_anis_over_2[2]) ; + else sumtheta += 0.5 * ODSLF_Theta_XXZ((new_tanhlambda - tanhlambda[k][beta])/(1.0 - new_tanhlambda * tanhlambda[k][beta]), + chain.Str_L[j], chain.Str_L[k], chain.par[j], chain.par[k], chain.ta_n_anis_over_2); + } + sumtheta *= 2.0; + if (chain.par[j] == 1) arg = chain.ta_n_anis_over_2[chain.Str_L[j]] * tan(0.5 * (PI * Ix2[j][alpha] + sumtheta)/chain.Nsites); + + else if (chain.par[j] == -1) arg = -tan(0.5 * (PI * Ix2[j][alpha] + sumtheta)/chain.Nsites)/chain.ta_n_anis_over_2[chain.Str_L[j]]; + + else JSCerror("Invalid parities in Iterate_BAE."); + + } + + //cout << "Rapidity blows up !\t" << lambda[j][alpha] << "\t" << new_lambda << endl; + } + + return(new_lambda); + } + + bool ODSLF_XXZ_Bethe_State::Check_Rapidities() + { + bool nonan = true; + + for (int j = 0; j < chain.Nstrings; ++j) + for (int alpha = 0; alpha < base[j]; ++alpha) nonan *= !is_nan(lambda[j][alpha]); + + return nonan; + } + + void ODSLF_XXZ_Bethe_State::Compute_Energy () + { + DP sum = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum += sin(chain.Str_L[j] * chain.anis) / (chain.par[j] * cosh(2.0 * lambda[j][alpha]) - cos(chain.Str_L[j] * chain.anis)); + } + } + + sum *= - chain.J * sin(chain.anis); + + E = sum; + + return; + } + + /* + void XXZ_Bethe_State::Compute_Momentum () + { + int sum_Ix2 = 0; + DP sum_M = 0.0; + + for (int j = 0; j < chain.Nstrings; ++j) { + sum_M += 0.5 * (1.0 + chain.par[j]) * base[j]; + for (int alpha = 0; alpha < base[j]; ++alpha) { + sum_Ix2 += Ix2[j][alpha]; + } + } + + iK = (chain.Nsites/2) * int(sum_M + 0.1) - (sum_Ix2/2); // + 0.1: for safety... + + while (iK >= chain.Nsites) iK -= chain.Nsites; + while (iK < 0) iK += chain.Nsites; + + K = PI * sum_M - PI * sum_Ix2/chain.Nsites; + + while (K >= 2.0*PI) K -= 2.0*PI; + while (K < 0.0) K += 2.0*PI; + + return; + } + */ + void ODSLF_XXZ_Bethe_State::Build_Reduced_Gaudin_Matrix (SQMat >& Gaudin_Red) + { + + if (Gaudin_Red.size() != base.Nraptot) JSCerror("Passing matrix of wrong size in Build_Reduced_Gaudin_Matrix."); + + int index_jalpha; + int index_kbeta; + + DP sum_hbar_XXZ = 0.0; + + DP sinzetasq = pow(sin(chain.anis), 2.0); + + (*this).Compute_sinhlambda(); + (*this).Compute_coshlambda(); + + index_jalpha = 0; + for (int j = 0; j < chain.Nstrings; ++j) { + for (int alpha = 0; alpha < base[j]; ++alpha) { + index_kbeta = 0; + for (int k = 0; k < chain.Nstrings; ++k) { + for (int beta = 0; beta < base[k]; ++beta) { + + if ((j == k) && (alpha == beta)) { + + sum_hbar_XXZ = 0.0; + + for (int kp = 0; kp < chain.Nstrings; ++kp) { + for (int betap = 0; betap < base[kp]; ++betap) { + if (!((j == kp) && (alpha == betap))) + sum_hbar_XXZ + += ODSLF_ddlambda_Theta_XXZ (lambda[j][alpha] - lambda[kp][betap], chain.Str_L[j], chain.Str_L[kp], chain.par[j], chain.par[kp], + chain.si_n_anis_over_2); + } + } + + Gaudin_Red[index_jalpha][index_kbeta] + = complex ( chain.Nsites * ODSLF_hbar_XXZ (lambda[j][alpha], chain.Str_L[j], chain.par[j], chain.si_n_anis_over_2) - sum_hbar_XXZ); + } + + else { + if ((chain.Str_L[j] == 1) && (chain.Str_L[k] == 1)) + Gaudin_Red[index_jalpha][index_kbeta] = + complex ((chain.par[j] * chain.par[k] == 1) + ? chain.si_n_anis_over_2[4]/(pow(sinhlambda[j][alpha] * coshlambda[k][beta] + - coshlambda[j][alpha] * sinhlambda[k][beta], 2.0) + sinzetasq) + : chain.si_n_anis_over_2[4]/(-pow(coshlambda[j][alpha] * coshlambda[k][beta] + - sinhlambda[j][alpha] * sinhlambda[k][beta], 2.0) + sinzetasq) ); + else + Gaudin_Red[index_jalpha][index_kbeta] = complex (ODSLF_ddlambda_Theta_XXZ (lambda[j][alpha] - lambda[k][beta], chain.Str_L[j], chain.Str_L[k], + chain.par[j], chain.par[k], chain.si_n_anis_over_2)); + } + index_kbeta++; + } + } + index_jalpha++; + } + } + return; + } + + // **************************************************************************************************** + + // non-member functions + + inline DP ODSLF_fbar_XXZ (DP tanhlambda, int par, DP tannzetaover2) + { + DP result = 0.0; + + if (par == 1) result = 2.0 * atan(tanhlambda/tannzetaover2); + + else if (par == -1) result = -2.0 * atan(tanhlambda * tannzetaover2); + + else JSCerror("Faulty parity in ODSLF_fbar_XXZ."); + + return (result); + } + + DP ODSLF_Theta_XXZ (DP tanhlambda, int nj, int nk, int parj, int park, DP* tannzetaover2) + { + DP result = 0.0; + + if ((nj == 1) && (nk == 1)) result = ODSLF_fbar_XXZ(tanhlambda, parj*park, tannzetaover2[2]); + + else { + + result = (nj == nk) ? 0.0 : ODSLF_fbar_XXZ(tanhlambda, parj*park, tannzetaover2[fabs(nj - nk)]); + + for (int a = 1; a < JSC::min(nj, nk); ++a) result += 2.0 * ODSLF_fbar_XXZ(tanhlambda, parj*park, tannzetaover2[fabs(nj - nk) + 2*a]); + + result += ODSLF_fbar_XXZ(tanhlambda, parj*park, tannzetaover2[nj + nk]); + } + + return (result); + } + + DP ODSLF_hbar_XXZ (DP lambda, int n, int par, DP* si_n_anis_over_2) + { + DP result = 0.0; + + if (par == 1) result = si_n_anis_over_2[2*n]/(pow(sinh(lambda), 2.0) + pow(si_n_anis_over_2[n], 2.0)); + + else if (par == -1) result = si_n_anis_over_2[2*n]/(-pow(cosh(lambda), 2.0) + pow(si_n_anis_over_2[n], 2.0)); + + else JSCerror("Faulty parity in ODSLF_hbar_XXZ."); + + return (result); + } + + DP ODSLF_ddlambda_Theta_XXZ (DP lambda, int nj, int nk, int parj, int park, DP* si_n_anis_over_2) + { + DP result = (nj == nk) ? 0.0 : ODSLF_hbar_XXZ(lambda, fabs(nj - nk), parj*park, si_n_anis_over_2); + + for (int a = 1; a < JSC::min(nj, nk); ++a) result += 2.0 * ODSLF_hbar_XXZ(lambda, fabs(nj - nk) + 2*a, parj*park, si_n_anis_over_2); + + result += ODSLF_hbar_XXZ(lambda, nj + nk, parj*park, si_n_anis_over_2); + + return (result); + } + + +} // namespace JSC diff --git a/src/ODSLF/ln_Smin_ME_ODSLF_XXZ.cc b/src/ODSLF/ln_Smin_ME_ODSLF_XXZ.cc new file mode 100644 index 0000000..dbf99ce --- /dev/null +++ b/src/ODSLF/ln_Smin_ME_ODSLF_XXZ.cc @@ -0,0 +1,326 @@ +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + +inline complex ln_Fn_F (ODSLF_XXZ_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + int arg = 0; + int absarg = 0; + int par_comb_1, par_comb_2; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + + par_comb_1 = B.chain.par[j] == B.chain.par[k] ? 1 : 0; + par_comb_2 = B.chain.par[k] == B.chain.par[j] ? 0 : B.chain.par[k]; + + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) { + + arg = B.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + absarg = abs(arg); + /* + prod_temp *= 0.5 * //done later... + ((B.sinhlambda[j][alpha] * B.coshlambda[k][beta] - B.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (B.chain.co_n_anis_over_2[absarg] * (1.0 + B.chain.par[j] * B.chain.par[k]) + - sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * (B.chain.par[k] - B.chain.par[j])) + + II * (B.coshlambda[j][alpha] * B.coshlambda[k][beta] - B.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * (1.0 + B.chain.par[j] * B.chain.par[k]) + + B.chain.co_n_anis_over_2[absarg] * (B.chain.par[k] - B.chain.par[j])) ); + */ + + prod_temp *= ((B.sinhlambda[j][alpha] * B.coshlambda[k][beta] - B.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (B.chain.co_n_anis_over_2[absarg] * par_comb_1 - sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * par_comb_2) + + II * (B.coshlambda[j][alpha] * B.coshlambda[k][beta] - B.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * par_comb_1 + B.chain.co_n_anis_over_2[absarg] * par_comb_2)); + } + + if (counter++ > 100) { // we do at most 100 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + + }}} + + return(ans + log(prod_temp)); +} + +inline complex ln_Fn_G (ODSLF_XXZ_Bethe_State& A, ODSLF_XXZ_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + int arg = 0; + int absarg = 0; + int par_comb_1, par_comb_2; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + + par_comb_1 = A.chain.par[j] == B.chain.par[k] ? 1 : 0; + par_comb_2 = B.chain.par[k] == A.chain.par[j] ? 0 : B.chain.par[k]; + + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + arg = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + absarg = abs(arg); + /* + prod_temp *= 0.5 * //done later... + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg] * (B.chain.par[k] - A.chain.par[j])) ); + */ + prod_temp *= ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg] * par_comb_1 - sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * par_comb_2) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * par_comb_1 + A.chain.co_n_anis_over_2[absarg] * par_comb_2)); + + if (counter++ > 100) { // we do at most 100 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + }}} + + return(ans + log(prod_temp)); +} + +inline complex Fn_K (ODSLF_XXZ_Bethe_State& A, int j, int alpha, int a, ODSLF_XXZ_Bethe_State& B, int k, int beta, int b) +{ + int arg1 = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + int absarg1 = abs(arg1); + int arg2 = arg1 + 2; + int absarg2 = abs(arg2); + + return(4.0/( + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg1] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg1) * A.chain.si_n_anis_over_2[absarg1] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg1) * A.chain.si_n_anis_over_2[absarg1] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg1] * (B.chain.par[k] - A.chain.par[j])) ) + * + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg2] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg2) * A.chain.si_n_anis_over_2[absarg2] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg2) * A.chain.si_n_anis_over_2[absarg2] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg2] * (B.chain.par[k] - A.chain.par[j])) ) + )); + +} + +inline complex Fn_L (ODSLF_XXZ_Bethe_State& A, int j, int alpha, int a, ODSLF_XXZ_Bethe_State& B, int k, int beta, int b) +{ + return (sinh(2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + + 0.25 * II * PI * complex(-A.chain.par[j] + B.chain.par[k]))) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + +complex ln_Smin_ME (ODSLF_XXZ_Bethe_State& A, ODSLF_XXZ_Bethe_State& B) +{ + // This function returns the natural log of the S^- operator matrix element. + // The A and B states can contain strings. + + // Check that the two states are compatible + + if (A.chain != B.chain) JSCerror("Incompatible ODSLF_XXZ_Chains in Smin matrix element."); + + // Check that A and B are Mdown-compatible: + + if (A.base.Mdown != B.base.Mdown + 1) JSCerror("Incompatible Mdown between the two states in Smin matrix element!"); + + // Compute the sinh and cosh of rapidities + + A.Compute_sinhlambda(); + A.Compute_coshlambda(); + B.Compute_sinhlambda(); + B.Compute_coshlambda(); + + // Some convenient arrays + + ODSLF_Lambda re_ln_Fn_F_B_0(B.chain, B.base); + ODSLF_Lambda im_ln_Fn_F_B_0(B.chain, B.base); + ODSLF_Lambda re_ln_Fn_G_0(B.chain, B.base); + ODSLF_Lambda im_ln_Fn_G_0(B.chain, B.base); + ODSLF_Lambda re_ln_Fn_G_2(B.chain, B.base); + ODSLF_Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(sinh(A.lambda[i][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - A.chain.par[i])))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - B.chain.par[i]))) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - B.chain.par[i])))); + + // Define the F ones earlier... + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + DP logabssinzeta = log(abs(sin(A.chain.anis))); + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) + ln_prod3 += ln_Fn_F(A, j, alpha, a - 1); // assume only one-strings here + + ln_prod3 -= A.base.Mdown * log(abs(sin(A.chain.anis))); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + ln_prod4 -= B.base.Mdown * log(abs(sin(B.chain.anis))); + + // Now proceed to build the Hm matrix + + SQMat_CX Hm(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + complex one_over_A_sinhlambda_sq_plus_sinzetaover2sq; + + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + one_over_A_sinhlambda_sq_plus_sinzetaover2sq = 1.0/((sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + * (sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + + pow(sin(0.5*A.chain.anis), 2.0)); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinzeta); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinzeta); + + Prod_powerN = pow( B.chain.par[k] == 1 ? + (B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + : + (B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + , complex (B.chain.Nsites)); + + Hm[index_a][index_b] = Fn_K_0_G_0 - (1.0 - 2.0 * (B.base.Mdown % 2)) * // MODIF from XXZ + Prod_powerN * Fn_K_1_G_2; + + } // if (B.chain.Str_L == 1) + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += (1.0 - 2.0 * (B.base.Mdown % 2)) * // MODIF from XXZ + Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + /* + sum2 = 0.0; + + for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + + */ + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]] + logabssinzeta); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_Fn_F(B, k, beta, jsum - 1)) + logabssinzeta); + // include all string contributions F_B_0 in this term + + Hm[index_a][index_b] = prod_num * sum1; + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + + // now define the elements Hm[a][M] + + Hm[index_a][B.base.Mdown] = one_over_A_sinhlambda_sq_plus_sinzetaover2sq; + + index_a++; + }}} // sums over j, alpha, a + + complex ln_ME_sq = log(1.0 * A.chain.Nsites) + real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) + + 2.0 * real(lndet_LU_CX_dstry(Hm)) + logabssinzeta - A.lnnorm - B.lnnorm; + + //return(ln_ME_sq); + return(0.5 * ln_ME_sq); // Return ME, not MEsq + +} + +} // namespace JSC diff --git a/src/ODSLF/ln_Sz_ME_ODSLF_XXZ.cc b/src/ODSLF/ln_Sz_ME_ODSLF_XXZ.cc new file mode 100644 index 0000000..971c2b2 --- /dev/null +++ b/src/ODSLF/ln_Sz_ME_ODSLF_XXZ.cc @@ -0,0 +1,330 @@ +#include "JSC.h" + +using namespace JSC; + +namespace JSC { + +inline complex ln_Fn_F (ODSLF_XXZ_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + int arg = 0; + int absarg = 0; + int par_comb_1, par_comb_2; + + for (int j = 0; j < B.chain.Nstrings; ++j) { + + par_comb_1 = B.chain.par[j] == B.chain.par[k] ? 1 : 0; + par_comb_2 = B.chain.par[k] == B.chain.par[j] ? 0 : B.chain.par[k]; + + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + + for (int a = 1; a <= B.chain.Str_L[j]; ++a) { + + if (!((j == k) && (alpha == beta) && (a == b))) { + + arg = B.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + absarg = abs(arg); + /* + prod_temp *= 0.5 * + ((B.sinhlambda[j][alpha] * B.coshlambda[k][beta] - B.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (B.chain.co_n_anis_over_2[absarg] * (1.0 + B.chain.par[j] * B.chain.par[k]) + - sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * (B.chain.par[k] - B.chain.par[j])) + + II * (B.coshlambda[j][alpha] * B.coshlambda[k][beta] - B.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) + * B.chain.si_n_anis_over_2[absarg] * (1.0 + B.chain.par[j] * B.chain.par[k]) + + B.chain.co_n_anis_over_2[absarg] * (B.chain.par[k] - B.chain.par[j])) ); + */ + prod_temp *= ((B.sinhlambda[j][alpha] * B.coshlambda[k][beta] - B.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (B.chain.co_n_anis_over_2[absarg] * par_comb_1 - sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * par_comb_2) + + II * (B.coshlambda[j][alpha] * B.coshlambda[k][beta] - B.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * B.chain.si_n_anis_over_2[absarg] * par_comb_1 + B.chain.co_n_anis_over_2[absarg] * par_comb_2)); + + } + + if (counter++ > 100) { // we do at most 100 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + + }}} + + return(ans + log(prod_temp)); +} + +inline complex ln_Fn_G (ODSLF_XXZ_Bethe_State& A, ODSLF_XXZ_Bethe_State& B, int k, int beta, int b) +{ + complex ans = 0.0; + complex prod_temp = 1.0; + int counter = 0; + int arg = 0; + int absarg = 0; + int par_comb_1, par_comb_2; + + for (int j = 0; j < A.chain.Nstrings; ++j) { + + par_comb_1 = A.chain.par[j] == B.chain.par[k] ? 1 : 0; + par_comb_2 = B.chain.par[k] == A.chain.par[j] ? 0 : B.chain.par[k]; + + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + arg = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + absarg = abs(arg); + + /* + prod_temp *= 0.5 * + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg] * (B.chain.par[k] - A.chain.par[j])) ); + */ + prod_temp *= ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg] * par_comb_1 - sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * par_comb_2) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg) * A.chain.si_n_anis_over_2[absarg] * par_comb_1 + A.chain.co_n_anis_over_2[absarg] * par_comb_2)); + + if (counter++ > 100) { // we do at most 100 products before taking a log + ans += log(prod_temp); + prod_temp = 1.0; + counter = 0; + } + }}} + + return(ans + log(prod_temp)); +} + +inline complex Fn_K (ODSLF_XXZ_Bethe_State& A, int j, int alpha, int a, ODSLF_XXZ_Bethe_State& B, int k, int beta, int b) +{ + int arg1 = A.chain.Str_L[j] - B.chain.Str_L[k] - 2 * (a - b); + int absarg1 = abs(arg1); + int arg2 = arg1 + 2; + int absarg2 = abs(arg2); + + return(4.0/( + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg1] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg1) * A.chain.si_n_anis_over_2[absarg1] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg1) * A.chain.si_n_anis_over_2[absarg1] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg1] * (B.chain.par[k] - A.chain.par[j])) ) + * + ((A.sinhlambda[j][alpha] * B.coshlambda[k][beta] - A.coshlambda[j][alpha] * B.sinhlambda[k][beta]) + * (A.chain.co_n_anis_over_2[absarg2] * (1.0 + A.chain.par[j] * B.chain.par[k]) + - sgn_int(arg2) * A.chain.si_n_anis_over_2[absarg2] * (B.chain.par[k] - A.chain.par[j])) + + II * (A.coshlambda[j][alpha] * B.coshlambda[k][beta] - A.sinhlambda[j][alpha] * B.sinhlambda[k][beta]) + * (sgn_int(arg2) * A.chain.si_n_anis_over_2[absarg2] * (1.0 + A.chain.par[j] * B.chain.par[k]) + + A.chain.co_n_anis_over_2[absarg2] * (B.chain.par[k] - A.chain.par[j])) ) + )); + +} + +inline complex Fn_L (ODSLF_XXZ_Bethe_State& A, int j, int alpha, int a, ODSLF_XXZ_Bethe_State& B, int k, int beta, int b) +{ + return (sinh(2.0 * (A.lambda[j][alpha] - B.lambda[k][beta] + + 0.5 * II * B.chain.anis * (A.chain.Str_L[j] - B.chain.Str_L[k] - 2.0 * (a - b - 0.5)) + + 0.25 * II * PI * complex(-A.chain.par[j] + B.chain.par[k]))) + * pow(Fn_K (A, j, alpha, a, B, k, beta, b), 2.0)); +} + +complex ln_Sz_ME (ODSLF_XXZ_Bethe_State& A, ODSLF_XXZ_Bethe_State& B) +{ + // This function returns the natural log of the S^z operator matrix element. + // The A and B states can contain strings. + + // Check that the two states refer to the same XXZ_Chain + + if (A.chain != B.chain) JSCerror("Incompatible ODSLF_XXZ_Chains in Sz matrix element."); + + // Check that A and B are compatible: same Mdown + + if (A.base.Mdown != B.base.Mdown) JSCerror("Incompatible Mdown between the two states in Sz matrix element!"); + + // Compute the sinh and cosh of rapidities + + A.Compute_sinhlambda(); + A.Compute_coshlambda(); + B.Compute_sinhlambda(); + B.Compute_coshlambda(); + + // Some convenient arrays + + ODSLF_Lambda re_ln_Fn_F_B_0(B.chain, B.base); + ODSLF_Lambda im_ln_Fn_F_B_0(B.chain, B.base); + ODSLF_Lambda re_ln_Fn_G_0(B.chain, B.base); + ODSLF_Lambda im_ln_Fn_G_0(B.chain, B.base); + ODSLF_Lambda re_ln_Fn_G_2(B.chain, B.base); + ODSLF_Lambda im_ln_Fn_G_2(B.chain, B.base); + + complex ln_prod1 = 0.0; + complex ln_prod2 = 0.0; + complex ln_prod3 = 0.0; + complex ln_prod4 = 0.0; + + for (int i = 0; i < A.chain.Nstrings; ++i) + for (int alpha = 0; alpha < A.base.Nrap[i]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[i]; ++a) + ln_prod1 += log(norm(sinh(A.lambda[i][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - A.chain.par[i])))); + + for (int i = 0; i < B.chain.Nstrings; ++i) + for (int alpha = 0; alpha < B.base.Nrap[i]; ++alpha) + for (int a = 1; a <= B.chain.Str_L[i]; ++a) + if (norm(sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - B.chain.par[i]))) > 100.0 * MACHINE_EPS_SQ) + ln_prod2 += log(norm(sinh(B.lambda[i][alpha] + 0.5 * II * B.chain.anis * (B.chain.Str_L[i] + 1.0 - 2.0 * a - 1.0) + + 0.25 * II * PI * (1.0 - B.chain.par[i])))); + + // Define the F ones earlier... + + for (int j = 0; j < B.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < B.base.Nrap[j]; ++alpha) { + re_ln_Fn_F_B_0[j][alpha] = real(ln_Fn_F(B, j, alpha, 0)); + im_ln_Fn_F_B_0[j][alpha] = imag(ln_Fn_F(B, j, alpha, 0)); + re_ln_Fn_G_0[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 0)); + im_ln_Fn_G_0[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 0)); + re_ln_Fn_G_2[j][alpha] = real(ln_Fn_G(A, B, j, alpha, 2)); + im_ln_Fn_G_2[j][alpha] = imag(ln_Fn_G(A, B, j, alpha, 2)); + } + } + + DP logabssinzeta = log(abs(sin(A.chain.anis))); + + // Define regularized products in prefactors + + for (int j = 0; j < A.chain.Nstrings; ++j) + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + ln_prod3 += ln_Fn_F (A, j, alpha, a - 1); + } + + ln_prod3 -= A.base.Mdown * log(abs(sin(A.chain.anis))); + + for (int k = 0; k < B.chain.Nstrings; ++k) + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + if (b == 1) ln_prod4 += re_ln_Fn_F_B_0[k][beta]; + else if (b > 1) ln_prod4 += ln_Fn_F(B, k, beta, b - 1); + } + + ln_prod4 -= B.base.Mdown * log(abs(sin(B.chain.anis))); + + // Now proceed to build the Hm2P matrix + + SQMat_CX Hm2P(0.0, A.base.Mdown); + + int index_a = 0; + int index_b = 0; + + complex sum1 = 0.0; + complex sum2 = 0.0; + complex prod_num = 0.0; + complex Fn_K_0_G_0 = 0.0; + complex Prod_powerN = 0.0; + complex Fn_K_1_G_2 = 0.0; + complex two_over_A_sinhlambda_sq_plus_sinzetaover2sq; + + + for (int j = 0; j < A.chain.Nstrings; ++j) { + for (int alpha = 0; alpha < A.base.Nrap[j]; ++alpha) { + for (int a = 1; a <= A.chain.Str_L[j]; ++a) { + + index_b = 0; + + two_over_A_sinhlambda_sq_plus_sinzetaover2sq = 2.0/((sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + * (sinh(A.lambda[j][alpha] + 0.5 * II * A.chain.anis * (A.chain.Str_L[j] + 1.0 - 2.0 * a) + + 0.25 * II * PI * (1.0 - A.chain.par[j]))) + + pow(sin(0.5*A.chain.anis), 2.0)); + + for (int k = 0; k < B.chain.Nstrings; ++k) { + for (int beta = 0; beta < B.base.Nrap[k]; ++beta) { + for (int b = 1; b <= B.chain.Str_L[k]; ++b) { + + if (B.chain.Str_L[k] == 1) { + + // use simplified code for one-string here: original form of Hm2P matrix + + Fn_K_0_G_0 = Fn_K (A, j, alpha, a, B, k, beta, 0) * + exp(re_ln_Fn_G_0[k][beta] + II * im_ln_Fn_G_0[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinzeta); + Fn_K_1_G_2 = Fn_K (A, j, alpha, a, B, k, beta, 1) * + exp(re_ln_Fn_G_2[k][beta] + II * im_ln_Fn_G_2[k][beta] - re_ln_Fn_F_B_0[k][beta] + logabssinzeta); + + Prod_powerN = pow( B.chain.par[k] == 1 ? + (B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.sinhlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.coshlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + : + (B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] + II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + /(B.coshlambda[k][beta] * B.chain.co_n_anis_over_2[1] - II * B.sinhlambda[k][beta] * B.chain.si_n_anis_over_2[1]) + , complex (B.chain.Nsites)); + + Hm2P[index_a][index_b] = Fn_K_0_G_0 - (-1.0 + 2.0 * (B.base.Mdown % 2)) * // MODIF from XXZ + Prod_powerN * Fn_K_1_G_2 - two_over_A_sinhlambda_sq_plus_sinzetaover2sq + * exp(II*im_ln_Fn_F_B_0[k][beta] + logabssinzeta); + + } + + else { + + if (b <= B.chain.Str_L[k] - 1) Hm2P[index_a][index_b] = Fn_K(A, j, alpha, a, B, k, beta, b); + else if (b == B.chain.Str_L[k]) { + + Vect_CX ln_FunctionF(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionF[i] = ln_Fn_F (B, k, beta, i); + + Vect_CX ln_FunctionG(B.chain.Str_L[k] + 2); + for (int i = 0; i < B.chain.Str_L[k] + 2; ++i) ln_FunctionG[i] = ln_Fn_G (A, B, k, beta, i); + + sum1 = 0.0; + + sum1 += Fn_K (A, j, alpha, a, B, k, beta, 0) * exp(ln_FunctionG[0] + ln_FunctionG[1] - ln_FunctionF[0] - ln_FunctionF[1]); + + sum1 += (-1.0 + 2.0 * (B.base.Mdown % 2)) * // MODIF from XXZ + Fn_K (A, j, alpha, a, B, k, beta, B.chain.Str_L[k]) + * exp(ln_FunctionG[B.chain.Str_L[k]] + ln_FunctionG[B.chain.Str_L[k] + 1] + - ln_FunctionF[B.chain.Str_L[k]] - ln_FunctionF[B.chain.Str_L[k] + 1]); + + for (int jsum = 1; jsum < B.chain.Str_L[k]; ++jsum) + + sum1 -= Fn_L (A, j, alpha, a, B, k, beta, jsum) * + exp(ln_FunctionG[jsum] + ln_FunctionG[jsum + 1] - ln_FunctionF[jsum] - ln_FunctionF[jsum + 1]); + + sum2 = 0.0; + + for (int jsum = 1; jsum <= B.chain.Str_L[k]; ++jsum) sum2 += exp(ln_FunctionG[jsum] - ln_FunctionF[jsum]); + + prod_num = exp(II * im_ln_Fn_F_B_0[k][beta] + ln_FunctionF[1] - ln_FunctionG[B.chain.Str_L[k]] + logabssinzeta); + + for (int jsum = 2; jsum <= B.chain.Str_L[k]; ++jsum) + prod_num *= exp(ln_FunctionG[jsum] - real(ln_FunctionF[jsum - 1]) + logabssinzeta); + // include all string contributions F_B_0 in this term + + Hm2P[index_a][index_b] = prod_num * (sum1 - sum2 * two_over_A_sinhlambda_sq_plus_sinzetaover2sq); + + } // else if (b == B.chain.Str_L[k]) + } // else + + index_b++; + }}} // sums over k, beta, b + index_a++; + }}} // sums over j, alpha, a + + DP re_ln_det = real(lndet_LU_CX_dstry(Hm2P)); + + complex ln_ME_sq = log(0.25 * A.chain.Nsites) + real(ln_prod1 - ln_prod2) - real(ln_prod3) + real(ln_prod4) + + 2.0 * /*real(lndet_LU_CX_dstry(Hm2P))*/ re_ln_det - A.lnnorm - B.lnnorm; + + //cout << endl << ln_prod1 << "\t" << ln_prod2 << "\t" << ln_prod3 << "\t" << ln_prod4 << "\t" << A.lnnorm << "\t" << B.lnnorm + // << "\t" << re_ln_det << "\t" << ln_ME_sq << endl; + + //return(ln_ME_sq); + return(0.5 * ln_ME_sq); // Return ME, not MEsq + +} + +} // namespace JSC diff --git a/src/RICHARDSON/Richardson.cc b/src/RICHARDSON/Richardson.cc new file mode 100644 index 0000000..4f3d13b --- /dev/null +++ b/src/RICHARDSON/Richardson.cc @@ -0,0 +1,210 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: src/RICHARDSON/Rischardson.cc + +Purpose: Richardson model + + +***********************************************************/ + + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + bool Solve_Richardson_Quad_S (DP g_int, const Vect_DP epsilon, const Vect_DP sumoneovereps, + const Vect_CX S, const Vect_CX sumSovereps, int j, complex& Sjleft, complex& Sjright) + { + // This solves the equation + // S^2_j - \sum_{i \neq j}^N \frac{S_j - S_i}{\epsilon_j - \epsilon_i} - (1/g) S_j = 0. + // for S_j, and puts the result in Sj. + // If the result is real, true is returned, otherwise false. + + //DP bquad = -1.0/g_int - sumoneovereps[j]; + //DP cquad = sumSovereps[j]; + /* + DP discr = 0.0; + bool retval = false; + //if ((discr = bquad * bquad - 4.0 * cquad) >= 0.0) { + if ((discr = (1.0/g_int + sumoneovereps[j]) * (1.0/g_int + sumoneovereps[j]) - 4.0 * sumSovereps[j]) >= 0.0) { + discr = sqrt(discr); + Sjleft = 0.5 * (-(1.0/g_int + sumoneovereps[j]) - discr); + Sjright = 0.5 * (-(1.0/g_int + sumoneovereps[j]) + discr); + retval = true; + } + + return(retval); + */ + + complex discr = sqrt((1.0/g_int + sumoneovereps[j]) * (1.0/g_int + sumoneovereps[j]) - 4.0 * sumSovereps[j]); + //complex discr = sqrt(JSC::max(0.0, real((1.0/g_int + sumoneovereps[j]) * (1.0/g_int + sumoneovereps[j]) - 4.0 * sumSovereps[j]))); + Sjleft = 0.5 * (1.0/g_int + sumoneovereps[j] - discr); + Sjright = 0.5 * (1.0/g_int + sumoneovereps[j] + discr); + return(true); + } + + bool Solve_Richardson_Quad_S (DP g_int, const Vect_DP epsilon, Vect_CX& S, Vect leftroot, DP req_prec) + { + // This solves the set of equations + // S^2_j - \sum_{i \neq j}^N \frac{S_j - S_i}{\epsilon_j - \epsilon_i} - (1/g) S_j = 0. + + int Nlevels = epsilon.size(); + Vect_DP absLHSRE (Nlevels); // absolute value of left-hand side of Richardson eqns for S_j written as LHSRE == 0. + + Vect_DP sumoneovereps (0.0, Nlevels); + Vect_CX sumSovereps (Nlevels); + + for (int j = 0; j < Nlevels; ++j) { + sumoneovereps[j] = 0.0; + for (int i = 0; i < Nlevels; ++i) if (i != j) sumoneovereps[j] += 1.0/(epsilon[j] - epsilon[i]); + sumSovereps[j] = 0.0; + for (int i = 0; i < Nlevels; ++i) if (i != j) sumSovereps[j] += S[i]/(epsilon[j] - epsilon[i]); + } + + DP maxabsLHSRE = 0.0; + int jmax = 0; + complex Sleft, Sright; + + DP damping = 1.0; + + do { + + // Calculate the deviations, identify the largest: + maxabsLHSRE = 0.0; + for (int j = 0; j < Nlevels; ++j) { + absLHSRE[j] = abs(S[j] * S[j] - S[j] * sumoneovereps[j] + sumSovereps[j] - S[j]/g_int); + if (absLHSRE[j] > maxabsLHSRE) { + maxabsLHSRE = absLHSRE[j]; + jmax = j; + } + } + + cout << "identified jmax = " << jmax << " with maxabsLHSRE = " << maxabsLHSRE << endl; + + cout << "S[jmax]: from " << S[jmax]; + + // Re-solve for jmax: + if (Solve_Richardson_Quad_S (g_int, epsilon, sumoneovereps, S, sumSovereps, jmax, Sleft, Sright)) { + //if (leftroot[jmax] == true) S[jmax] = damping * Sleft + (1.0 - damping) * S[jmax]; + //else S[jmax] = damping * Sright + (1.0 - damping) * S[jmax]; + if (abs(S[jmax] - Sleft) < abs(S[jmax] - Sright)) S[jmax] = damping * Sleft + (1.0 - damping) * S[jmax]; + else S[jmax] = damping * Sright + (1.0 - damping) * S[jmax]; + } + else { + JSCerror("Complex jmax root."); + } + cout << " to " << S[jmax] << endl; + + // Re-solve also for a random j, given by + //int jrand = int(maxabsLHSRE * 1.0e+6) % Nlevels; + int jrand = rand() % Nlevels; + cout << "identified jrand = " << jrand << endl; + cout << "S[jrand]: from " << S[jrand]; + + if (Solve_Richardson_Quad_S (g_int, epsilon, sumoneovereps, S, sumSovereps, jrand, Sleft, Sright)) { + //if (leftroot[jrand] == true) S[jrand] = damping * Sleft + (1.0 - damping) * S[jrand]; + //else S[jrand] = damping * Sright + (1.0 - damping) * S[jrand]; + if (abs(S[jrand] - Sleft) < abs(S[jrand] - Sright)) S[jrand] = damping * Sleft + (1.0 - damping) * S[jrand]; + else S[jrand] = damping * Sright + (1.0 - damping) * S[jrand]; + } + else { + JSCerror("Complex jrand root."); + } + cout << " to " << S[jrand] << endl; + + /* + for (int j = 0; j < Nlevels; ++j) { + if (Solve_Richardson_Quad_S (g_int, epsilon, sumoneovereps, S, sumSovereps, j, Sleft, Sright)) { + if (leftroot[j] == true) S[j] = damping * Sleft + (1.0 - damping) * S[j]; + else S[j] = damping * Sright + (1.0 - damping) * S[j]; + } + else { + JSCerror("Complex root."); + } + } + */ + + // Recalculate all sums: + for (int j = 0; j < Nlevels; ++j) { + sumSovereps[j] = 0.0; + for (int i = 0; i < Nlevels; ++i) if (i != j) sumSovereps[j] += S[i]/(epsilon[j] - epsilon[i]); + } + + } while (maxabsLHSRE > req_prec); + + return(true); + } + + + + +} // namespace JSC + + + +int main () +{ + DP g_int = 0.1; + + int Nlevels = 10; + + Vect_DP epsilon (Nlevels); + for (int i = 0; i < Nlevels; ++i) epsilon[i] = -0.5 * (Nlevels - 1.0) + i; + + DP req_prec = 1.0e-10; + + Vect_CX S (2.0, Nlevels); + + // Initial conditions: + /* + // Sa252 g = 0.78 + S[0] = -0.6142; + S[1] = -0.7344; + S[2] = -0.9177; + S[3] = -1.2414; + S[4] = -2.06103; + S[5] = 3.06103; + S[6] = 2.2414; + S[7] = 1.9178; + S[8] = 1.7344; + S[9] = 1.6142; + //for (int j = 0; j < Nlevels; ++j) S[j] /= g_int; + */ + + + Vect leftroot (true, Nlevels); + + //for (int j = Nlevels/2; j < Nlevels; ++j) leftroot[j] = true; + + leftroot[0] = 0; + leftroot[1] = 1; + leftroot[2] = 0; + leftroot[3] = 1; + leftroot[4] = 0; + leftroot[5] = 1; + leftroot[6] = 1; + leftroot[7] = 1; + leftroot[8] = 1; + leftroot[9] = 1; + + //for (int j = 0; j < Nlevels; ++j) leftroot[j] = !leftroot[j]; + + cout << Solve_Richardson_Quad_S (g_int, epsilon, S, leftroot, req_prec) << endl; + cout << leftroot << endl; + for (int j = 0; j < Nlevels; ++j) S[j] *= g_int; + cout << S << endl; + + return(0); +} + + + diff --git a/src/SCAN/Descendents.cc b/src/SCAN/Descendents.cc new file mode 100644 index 0000000..0cc3dd2 --- /dev/null +++ b/src/SCAN/Descendents.cc @@ -0,0 +1,1805 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Descendents.cc + +Purpose: New implementation of state scanning: + functions to descend down hierarchy of states. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + + Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2, const Vect& Ix2_min, const Vect& Ix2_max, bool disperse_only_current_exc, bool preserve_nexc) + { + //cout << "Up start" << endl; + // Given an OriginIx2 and a BaseScanIx2 (by which we mean a set of Ix2 providing a basis for scanning), + // this function returns a vector of the labels of the states obtained by each allowable + // one-step increase of the quantum numbers. + + // Rules for moving quantum numbers: + // - quantum numbers preserve their ordering (an Umklapp is thus made of many one-step movements) + // - if a quantum number has moved as compared to BaseScanIx2, then it can only move further in the same direction + + // The ordering rule for allowable increases is that we start from the topmost, rightmost quantum number. + // The increases can thus be applied to quantum numbers below the bottom-most, left-most right-displaced quantum nr. + + // The preserve_nexc boolean flags whether we maintain or increase the nr of ph pairs. + // This is only checked if disperse_only_current_exc == false (otherwise it's irrelevant). + + Vect > ScanIx2 = Return_Ix2_from_Label (ScanIx2_label, OriginIx2); + + // Determine the level and index of the bottom-most left-most right-moving quantum number sits: + int exclevel = -1; + int excindex = 0; + bool excfound = false; + do { + exclevel++; + if (exclevel == ScanIx2.size()) { // there isn't a single right-moving quantum number in ScanIx2 + break; + } + for (int alpha = 0; alpha < ScanIx2[exclevel].size(); ++alpha) + if (ScanIx2[exclevel][alpha] > BaseScanIx2[exclevel][alpha]) { + excindex = alpha; + excfound = true; + break; + } + } while (!excfound); + // If we haven't found an excitation, then exclevel == ScanIx2.size() and excindex = 0; + + //cout << "exclevel = " << exclevel << endl; + //cout << "excindex = " << excindex << endl; + + // The quantum numbers which we can move right are thus those + // with (j < exclevel) and (j == exclelev and alpha <= excindex) + + int ndesc_possible = 1; + if (!disperse_only_current_exc) { + ndesc_possible = 0; + for (int j = 0; j <= JSC::min(exclevel, ScanIx2.size() -1); ++j) ndesc_possible += ScanIx2[j].size(); + } + + // Now construct the actual descendents: + Vect desclabelfound (ndesc_possible); + int ndesc_found = 0; + + if (!disperse_only_current_exc && !preserve_nexc) + // List descendents with a new excitation at a lower level: + // this can only create an additional ph pair + for (int j = 0; j < exclevel; ++j) + for (int alpha = 0; alpha < ScanIx2[j].size(); ++alpha) { + if (ScanIx2[j][alpha] >= BaseScanIx2[j][alpha] // candidate either undisplaced or right-moving + && !ScanIx2[j].includes(ScanIx2[j][alpha] + 2) // there is a hole to its right + && ScanIx2[j][alpha] + 2 <= Ix2_max[j]) { // this new quantum number fits within the limits + // We've found a descendent + ScanIx2[j][alpha] += 2; + desclabelfound[ndesc_found] = Return_State_Label (ScanIx2, OriginIx2); + ScanIx2[j][alpha] -= 2; + ndesc_found++; + } + } + + // List descendents with a new excitation at exclevel: + if (exclevel < ScanIx2.size()) { // excfound == true, excindex is now guaranteed < ScanIx2[exclevel].size() + int alphamin = (disperse_only_current_exc ? excindex : 0); + int alphamax = (disperse_only_current_exc ? excindex : excindex - 1); + //for (int alpha = 0; alpha <= excindex; ++alpha) { + for (int alpha = alphamin; alpha <= alphamax; ++alpha) { + if (ScanIx2[exclevel][alpha] >= BaseScanIx2[exclevel][alpha] + && !ScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] + 2) + && ScanIx2[exclevel][alpha] + 2 <= Ix2_max[exclevel]) { + // We've found a descendent + ScanIx2[exclevel][alpha] += 2; + desclabelfound[ndesc_found] = Return_State_Label (ScanIx2, OriginIx2); + ScanIx2[exclevel][alpha] -= 2; + // If we're dispersing a subleading Ix2, check whether we match the preserve_nexc condition: + // The descendent is acceptable if disperse_only_current_exc == true, + // or if preserve_nexc == true and nexc labels match, + // or if preserve_nexc == false and nexc labels don't match: + //if (!disperse_only_current_exc && (BaseScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] + 2) != (Extract_nexc_Label(desclabelfound[ndesc_found]).compare(Extract_nexc_Label(ScanIx2_label)) == 0))) JSCerror("Inconsistency in check in Descendents"); + if (disperse_only_current_exc + //|| (preserve_nexc == BaseScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] + 2))) + || (preserve_nexc == (Extract_nexc_Label(desclabelfound[ndesc_found]).compare(Extract_nexc_Label(ScanIx2_label)) == 0))) + ndesc_found++; + } + } + } //if (exclevel < ScanIx2.size()) + + // Resize desc: + Vect desclabelfound_resized(ndesc_found); + for (int idesc = 0; idesc < ndesc_found; ++idesc) + desclabelfound_resized[idesc] = desclabelfound[idesc]; + + //cout << "Up done" << endl; + + //return(descIx2found); + //return(desclabelfound); + return(desclabelfound_resized); + } + + + Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2, const Vect& Ix2_min, const Vect& Ix2_max, bool disperse_only_current_exc, bool preserve_nexc) + { + //cout << "Down start" << endl; + // Given an OriginIx2 and a BaseScanIx2 (by which we mean a set of Ix2 providing a basis for scanning), + // this function returns a vector of the labels of the states obtained by each allowable + // one-step decrease of the quantum numbers. + + // Rules for moving quantum numbers: + // - quantum numbers preserve their ordering (an Umklapp is thus made of many one-step movements) + // - if a quantum number has moved as compared to BaseScanIx2, then it can only move further in the same direction + + // The ordering rule for allowable decreases is that we start from the topmost, leftmost quantum number. + // The decreases can thus be applied to quantum numbers below the bottom-most, right-most left-displaced quantum nr. + + // The preserve_nexc boolean flags whether we maintain or increase the nr of ph pairs. + // This is only checked if disperse_only_current_exc == false (otherwise it's irrelevant). + + Vect > ScanIx2 = Return_Ix2_from_Label (ScanIx2_label, OriginIx2); + + // Determine the level and index of the bottom-most right-most left-moving quantum number sits: + int exclevel = -1; + int excindex = 0; + bool excfound = false; + + //cout << "Looking for exclevel and excindex for " << endl << "\tBaseIx2 = " << BaseScanIx2 << endl << "\tScanIx2 = " << ScanIx2 << endl; + do { + exclevel++; + if (exclevel == ScanIx2.size()) { // there isn't a single left-moving quantum number in ScanIx2 + break; + } + for (int alpha = ScanIx2[exclevel].size() - 1; alpha >= 0; --alpha) { + //cout << exclevel << "\t" << alpha << "\t" << ScanIx2[exclevel][alpha] << "\t" << BaseScanIx2[exclevel][alpha] << "\t" << (ScanIx2[exclevel][alpha] < BaseScanIx2[exclevel][alpha]) << endl; + if (ScanIx2[exclevel][alpha] < BaseScanIx2[exclevel][alpha]) { + excindex = alpha; + excfound = true; + break; + } + } + } while (!excfound); + // If we haven't found an excitation, then exclevel == ScanIx2.size() and excindex = 0; + //if (!excfound) excindex = ScanIx2[exclevel].size() - 1; + + //cout << "exclevel = " << exclevel << endl; + //cout << "excindex = " << excindex << endl; + + // The quantum numbers which we can move left are thus those + // with (j < exclevel) and (j == exclelev and alpha >= excindex) + + + int ndesc_possible = 1; + if (!disperse_only_current_exc) { + ndesc_possible = 0; + for (int j = 0; j <= JSC::min(exclevel, ScanIx2.size() - 1); ++j) ndesc_possible += ScanIx2[j].size(); + } + + // Now construct the actual descendents: + Vect desclabelfound (ndesc_possible); + int ndesc_found = 0; + + if (!disperse_only_current_exc && !preserve_nexc) + // List descendents with a new excitation at a lower level: + for (int j = 0; j < exclevel; ++j) + for (int alpha = 0; alpha < ScanIx2[j].size(); ++alpha) { + if (ScanIx2[j][alpha] <= BaseScanIx2[j][alpha] // candidate either undisplaced or left-moving + && !ScanIx2[j].includes(ScanIx2[j][alpha] - 2) // there is a hole to its left + && ScanIx2[j][alpha] - 2 >= Ix2_min[j]) { // this new quantum number fits within the limits + // We've found a descendent + ScanIx2[j][alpha] -= 2; + desclabelfound[ndesc_found] = Return_State_Label (ScanIx2, OriginIx2); + ScanIx2[j][alpha] += 2; + ndesc_found++; + } + } + + // List descendents with a new excitation at exclevel: + if (exclevel < ScanIx2.size()) { // excfound == true, excindex is now guaranteed < ScanIx2[exclevel].size() + int alphamin = (disperse_only_current_exc ? excindex : excindex + 1); + int alphamax = (disperse_only_current_exc ? excindex : ScanIx2[exclevel].size() - 1); + //for (int alpha = ScanIx2[exclevel].size() - 1; alpha >= excindex; --alpha) { + for (int alpha = alphamax; alpha >= alphamin; --alpha) { + if (ScanIx2[exclevel][alpha] <= BaseScanIx2[exclevel][alpha] + && !ScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] - 2) + && ScanIx2[exclevel][alpha] - 2 >= Ix2_min[exclevel]) { + ScanIx2[exclevel][alpha] -= 2; + desclabelfound[ndesc_found] = Return_State_Label (ScanIx2, OriginIx2); + ScanIx2[exclevel][alpha] += 2; + //cout << "\tTesting .compare: " << Extract_nexc_Label(desclabelfound[ndesc_found]) << "\t" << Extract_nexc_Label(ScanIx2_label) << "\t" << (Extract_nexc_Label(desclabelfound[ndesc_found]).compare(Extract_nexc_Label(ScanIx2_label)) == 0) << "\tpreserve_nexc: " << preserve_nexc << endl; + if (disperse_only_current_exc + //|| (preserve_nexc == BaseScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] - 2))) + || (preserve_nexc == (Extract_nexc_Label(desclabelfound[ndesc_found]).compare(Extract_nexc_Label(ScanIx2_label)) == 0))) + //|| (preserve_nexc == strcmp(Extract_nexc_Label(desclabelfound[ndesc_found]), Extract_nexc_Label(ScanIx2_label)))) + ndesc_found++; + } + } + } // if (exclevel < ScanIx2.size()) + + // Resize desc: + Vect desclabelfound_resized(ndesc_found); + for (int idesc = 0; idesc < ndesc_found; ++idesc) + desclabelfound_resized[idesc] = desclabelfound[idesc]; + + //cout << "Down done" << endl; + //return(descIx2found); + //return(desclabelfound); + return(desclabelfound_resized); + } + + + Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2, const Vect& Ix2_min, const Vect& Ix2_max, bool disperse_only_current_exc_up, bool preserve_nexc_up, bool disperse_only_current_exc_down, bool preserve_nexc_down) + { + // Returns the labels of all states which are obtained from ScanIx2 by stepping p one step, and down one step. + Vect labels_up = Descendent_States_with_iK_Stepped_Up (ScanIx2_label, OriginIx2, BaseScanIx2, Ix2_min, Ix2_max, disperse_only_current_exc_up, preserve_nexc_up); + //cout << "labels_up = " << labels_up << endl; + + Vect labels_found(0); + for (int i = 0; i < labels_up.size(); ++i) { + labels_found.Append (Descendent_States_with_iK_Stepped_Down (labels_up[i], OriginIx2, BaseScanIx2, Ix2_min, Ix2_max, disperse_only_current_exc_down, preserve_nexc_down)); + } + + return(labels_found); + } + + + + // For symmetric state descending: + + Vect Descendent_States_with_iK_Stepped_Up_rightIx2only + (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2, + const Vect& Ix2_min, const Vect& Ix2_max, bool disperse_only_current_exc, bool preserve_nexc) + { + // This function raises an Ix2 in the right half of ScanIx2, and does the symmetric lowering in the lower half. + + // ASSUMPTIONS: OriginIx2 is a symmetric state. + + //cout << "Desc Up A" << endl; + + Vect > ScanIx2 = Return_Ix2_from_Label (ScanIx2_label, OriginIx2); + + //cout << "Desc Up B" << endl; + + // Determine the level and index of the bottom-most left-most right-moving quantum number sits: + int exclevel = -1; + int excindex = 0; + bool excfound = false; + do { + exclevel++; + if (exclevel == ScanIx2.size()) { // there isn't a single right-moving quantum number in ScanIx2 + break; + } + for (int alpha = (ScanIx2[exclevel].size() + 1)/2; alpha < ScanIx2[exclevel].size(); ++alpha) + if (ScanIx2[exclevel][alpha] > BaseScanIx2[exclevel][alpha]) { + excindex = alpha; + excfound = true; + break; + } + } while (!excfound); + // If we haven't found an excitation, then exclevel == ScanIx2.size() and excindex = 0; + + //cout << "exclevel = " << exclevel << "\t" << "excindex = " << excindex << endl; + + // The quantum numbers which we can move right are thus those + // with (j < exclevel) and (j == exclelev and alpha <= excindex) + + int ndesc_possible = 1; + if (!disperse_only_current_exc) { + ndesc_possible = 0; + for (int j = 0; j <= JSC::min(exclevel, ScanIx2.size() -1); ++j) ndesc_possible += ScanIx2[j].size(); + } + + // Now construct the actual descendents: + Vect desclabelfound (ndesc_possible); + int ndesc_found = 0; + + if (!disperse_only_current_exc && !preserve_nexc) + // List descendents with a new excitation at a lower level: + // this can only create an additional ph pair + for (int j = 0; j < exclevel; ++j) + for (int alpha = (ScanIx2[j].size() + 1)/2; alpha < ScanIx2[j].size(); ++alpha) { + if (ScanIx2[j][alpha] >= BaseScanIx2[j][alpha] // candidate either undisplaced or right-moving + && !ScanIx2[j].includes(ScanIx2[j][alpha] + 2) // there is a hole to its right + && ScanIx2[j][alpha] + 2 <= Ix2_max[j]) { // this new quantum number fits within the limits + // We've found a descendent + ScanIx2[j][alpha] += 2; + ScanIx2[j][ScanIx2[j].size() - 1 - alpha] -= 2; + desclabelfound[ndesc_found] = Return_State_Label (ScanIx2, OriginIx2); + ScanIx2[j][alpha] -= 2; + ScanIx2[j][ScanIx2[j].size() - 1 - alpha] += 2; + ndesc_found++; + } + } + + // List descendents with a new excitation at exclevel: + if (exclevel < ScanIx2.size()) { // excfound == true, excindex is now guaranteed < ScanIx2[exclevel].size() + int alphamin = (disperse_only_current_exc ? excindex : 0); + alphamin = JSC::max((ScanIx2[exclevel].size() + 1)/2, alphamin); + int alphamax = (disperse_only_current_exc ? excindex : excindex - 1); + //for (int alpha = 0; alpha <= excindex; ++alpha) { + //cout << "alphamin = " << alphamin << "\talphamax = " << alphamax << endl; + for (int alpha = alphamin; alpha <= alphamax; ++alpha) { + if (ScanIx2[exclevel][alpha] >= BaseScanIx2[exclevel][alpha] + && !ScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] + 2) + && ScanIx2[exclevel][alpha] + 2 <= Ix2_max[exclevel]) { + // We've found a descendent + ScanIx2[exclevel][alpha] += 2; + ScanIx2[exclevel][ScanIx2[exclevel].size() - 1 - alpha] -= 2; + //cout << "Desc Up a" << endl; + //cout << "ScanIx2[exclevel] = " << ScanIx2[exclevel] << endl; + desclabelfound[ndesc_found] = Return_State_Label (ScanIx2, OriginIx2); + //cout << "Desc Up b" << endl; + ScanIx2[exclevel][alpha] -= 2; + ScanIx2[exclevel][ScanIx2[exclevel].size() - 1 - alpha] += 2; + // If we're dispersing a subleading Ix2, check whether we match the preserve_nexc condition: + // The descendent is acceptable if disperse_only_current_exc == true, + // or if preserve_nexc == true and nexc labels match, + // or if preserve_nexc == false and nexc labels don't match: + //if (!disperse_only_current_exc && (BaseScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] + 2) != (Extract_nexc_Label(desclabelfound[ndesc_found]).compare(Extract_nexc_Label(ScanIx2_label)) == 0))) JSCerror("Inconsistency in check in Descendents"); + if (disperse_only_current_exc + //|| (preserve_nexc == BaseScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] + 2))) + || (preserve_nexc == (Extract_nexc_Label(desclabelfound[ndesc_found]).compare(Extract_nexc_Label(ScanIx2_label)) == 0))) + ndesc_found++; + } + } + } //if (exclevel < ScanIx2.size()) + + //cout << "Desc Up D" << endl; + + // Resize desc: + Vect desclabelfound_resized(ndesc_found); + for (int idesc = 0; idesc < ndesc_found; ++idesc) + desclabelfound_resized[idesc] = desclabelfound[idesc]; + + //cout << "Up done" << endl; + + //return(descIx2found); + //return(desclabelfound); + return(desclabelfound_resized); + } + + + Vect Descendent_States_with_iK_Stepped_Down_rightIx2only + (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2, + const Vect& Ix2_min, const Vect& Ix2_max, bool disperse_only_current_exc, bool preserve_nexc) + { + Vect > ScanIx2 = Return_Ix2_from_Label (ScanIx2_label, OriginIx2); + + // Determine the level and index of the bottom-most right-most left-moving quantum number sits: + int exclevel = -1; + int excindex = 0; + bool excfound = false; + + //cout << "Looking for exclevel and excindex for " << endl << "\tBaseIx2 = " << BaseScanIx2 << endl << "\tScanIx2 = " << ScanIx2 << endl; + do { + exclevel++; + if (exclevel == ScanIx2.size()) { // there isn't a single left-moving quantum number in ScanIx2 + break; + } + for (int alpha = ScanIx2[exclevel].size() - 1; alpha >= (ScanIx2[exclevel].size() + 1)/2 ; --alpha) { + //cout << exclevel << "\t" << alpha << "\t" << ScanIx2[exclevel][alpha] << "\t" << BaseScanIx2[exclevel][alpha] << "\t" << (ScanIx2[exclevel][alpha] < BaseScanIx2[exclevel][alpha]) << endl; + if (ScanIx2[exclevel][alpha] < BaseScanIx2[exclevel][alpha]) { + excindex = alpha; + excfound = true; + break; + } + } + } while (!excfound); + // If we haven't found an excitation, then exclevel == ScanIx2.size() and excindex = 0; + //if (!excfound) excindex = ScanIx2[exclevel].size() - 1; + + //cout << "exclevel = " << exclevel << endl; + //cout << "excindex = " << excindex << endl; + + // The quantum numbers which we can move left are thus those + // with (j < exclevel) and (j == exclevel and alpha >= excindex) + + + int ndesc_possible = 1; + if (!disperse_only_current_exc) { + ndesc_possible = 0; + for (int j = 0; j <= JSC::min(exclevel, ScanIx2.size() - 1); ++j) ndesc_possible += ScanIx2[j].size(); + } + + // Now construct the actual descendents: + Vect desclabelfound (ndesc_possible); + int ndesc_found = 0; + + if (!disperse_only_current_exc && !preserve_nexc) + // List descendents with a new excitation at a lower level: + for (int j = 0; j < exclevel; ++j) + for (int alpha = (ScanIx2[j].size() + 1)/2; alpha < ScanIx2[j].size(); ++alpha) { + if (ScanIx2[j][alpha] <= BaseScanIx2[j][alpha] // candidate either undisplaced or left-moving + && !ScanIx2[j].includes(ScanIx2[j][alpha] - 2) // there is a hole to its left + && ScanIx2[j][alpha] - 2 >= Ix2_min[j]) { // this new quantum number fits within the limits + // We've found a descendent + ScanIx2[j][alpha] -= 2; + ScanIx2[j][ScanIx2[j].size() - 1 - alpha] += 2; + desclabelfound[ndesc_found] = Return_State_Label (ScanIx2, OriginIx2); + ScanIx2[j][alpha] += 2; + ScanIx2[j][ScanIx2[j].size() - 1 - alpha] -= 2; + ndesc_found++; + } + } + + // List descendents with a new excitation at exclevel: + if (exclevel < ScanIx2.size()) { // excfound == true, excindex is now guaranteed < ScanIx2[exclevel].size() + int alphamin = (disperse_only_current_exc ? excindex : excindex + 1); + alphamin = JSC::max((ScanIx2[exclevel].size() + 1)/2, alphamin); + int alphamax = (disperse_only_current_exc ? excindex : ScanIx2[exclevel].size() - 1); + //for (int alpha = ScanIx2[exclevel].size() - 1; alpha >= excindex; --alpha) { + for (int alpha = alphamax; alpha >= alphamin; --alpha) { + if (ScanIx2[exclevel][alpha] <= BaseScanIx2[exclevel][alpha] + && !ScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] - 2) + && ScanIx2[exclevel][alpha] - 2 >= Ix2_min[exclevel]) { + ScanIx2[exclevel][alpha] -= 2; + ScanIx2[exclevel][ScanIx2[exclevel].size() - 1 - alpha] += 2; + desclabelfound[ndesc_found] = Return_State_Label (ScanIx2, OriginIx2); + ScanIx2[exclevel][alpha] += 2; + ScanIx2[exclevel][ScanIx2[exclevel].size() - 1 - alpha] -= 2; + //cout << "\tTesting .compare: " << Extract_nexc_Label(desclabelfound[ndesc_found]) << "\t" << Extract_nexc_Label(ScanIx2_label) << "\t" << (Extract_nexc_Label(desclabelfound[ndesc_found]).compare(Extract_nexc_Label(ScanIx2_label)) == 0) << "\tpreserve_nexc: " << preserve_nexc << endl; + if (disperse_only_current_exc + //|| (preserve_nexc == BaseScanIx2[exclevel].includes(ScanIx2[exclevel][alpha] - 2))) + || (preserve_nexc == (Extract_nexc_Label(desclabelfound[ndesc_found]).compare(Extract_nexc_Label(ScanIx2_label)) == 0))) + //|| (preserve_nexc == strcmp(Extract_nexc_Label(desclabelfound[ndesc_found]), Extract_nexc_Label(ScanIx2_label)))) + ndesc_found++; + } + } + } // if (exclevel < ScanIx2.size()) + + // Resize desc: + Vect desclabelfound_resized(ndesc_found); + for (int idesc = 0; idesc < ndesc_found; ++idesc) + desclabelfound_resized[idesc] = desclabelfound[idesc]; + + //cout << "Down done" << endl; + //return(descIx2found); + //return(desclabelfound); + return(desclabelfound_resized); + } + + + + // Specialization for Lieb-Liniger case: + //Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const Vect& OriginIx2, const Vect& BaseScanIx2) + Vect Descendent_States_with_iK_Stepped_Up (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc) + { + //Vect > ScanIx2here(1); + //ScanIx2here[0] = ScanIx2; + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginState.Ix2; + Vect > BaseScanIx2here(1); + //BaseScanIx2here[0] = BaseScanIx2; + BaseScanIx2here[0] = OriginState.Ix2; + Vect Ix2_min(1); + Ix2_min[0] = LIEBLIN_Ix2_MIN; + Vect Ix2_max(1); + Ix2_max[0] = LIEBLIN_Ix2_MAX; + + return (Descendent_States_with_iK_Stepped_Up (ScanIx2_label, OriginIx2here, BaseScanIx2here, Ix2_min, Ix2_max, disperse_only_current_exc, preserve_nexc)); + } + // Specialization for Lieb-Liniger case: + //Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const Vect& OriginIx2, const Vect& BaseScanIx2) + Vect Descendent_States_with_iK_Stepped_Down (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc) + { + //Vect > ScanIx2here(1); + //ScanIx2here[0] = ScanIx2; + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginState.Ix2; + Vect > BaseScanIx2here(1); + //BaseScanIx2here[0] = BaseScanIx2; + BaseScanIx2here[0] = OriginState.Ix2; + Vect Ix2_min(1); + Ix2_min[0] = LIEBLIN_Ix2_MIN; + Vect Ix2_max(1); + Ix2_max[0] = LIEBLIN_Ix2_MAX; + + return (Descendent_States_with_iK_Stepped_Down (ScanIx2_label, OriginIx2here, BaseScanIx2here, Ix2_min, Ix2_max, disperse_only_current_exc, preserve_nexc)); + } + + // Specialization for Lieb-Liniger case: + //Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const Vect& OriginIx2, const Vect& BaseScanIx2) + Vect Descendent_States_with_iK_Preserved (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc_up, bool preserve_nexc_up, bool disperse_only_current_exc_down, bool preserve_nexc_down) + { + //Vect > ScanIx2here(1); + //ScanIx2here[0] = ScanIx2; + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginState.Ix2; + Vect > BaseScanIx2here(1); + //BaseScanIx2here[0] = BaseScanIx2; + BaseScanIx2here[0] = OriginState.Ix2; + Vect Ix2_min(1); + Ix2_min[0] = LIEBLIN_Ix2_MIN; + Vect Ix2_max(1); + Ix2_max[0] = LIEBLIN_Ix2_MAX; + + return (Descendent_States_with_iK_Preserved (ScanIx2_label, OriginIx2here, BaseScanIx2here, Ix2_min, Ix2_max, disperse_only_current_exc_up, preserve_nexc_up, disperse_only_current_exc_down, preserve_nexc_down)); + } + + // Specialization for Lieb-Liniger case: + Vect Descendent_States_with_iK_Stepped_Up_rightIx2only + (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc) + { + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginState.Ix2; + Vect > BaseScanIx2here(1); + //BaseScanIx2here[0] = BaseScanIx2; + BaseScanIx2here[0] = OriginState.Ix2; + Vect Ix2_min(1); + Ix2_min[0] = LIEBLIN_Ix2_MIN; + Vect Ix2_max(1); + Ix2_max[0] = LIEBLIN_Ix2_MAX; + + return(Descendent_States_with_iK_Stepped_Up_rightIx2only (ScanIx2_label, OriginIx2here, BaseScanIx2here, Ix2_min, Ix2_max, disperse_only_current_exc, preserve_nexc)); + } + + Vect Descendent_States_with_iK_Stepped_Down_rightIx2only + (string ScanIx2_label, const LiebLin_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc) + { + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginState.Ix2; + Vect > BaseScanIx2here(1); + //BaseScanIx2here[0] = BaseScanIx2; + BaseScanIx2here[0] = OriginState.Ix2; + Vect Ix2_min(1); + Ix2_min[0] = LIEBLIN_Ix2_MIN; + Vect Ix2_max(1); + Ix2_max[0] = LIEBLIN_Ix2_MAX; + + return(Descendent_States_with_iK_Stepped_Down_rightIx2only (ScanIx2_label, OriginIx2here, BaseScanIx2here, Ix2_min, Ix2_max, disperse_only_current_exc, preserve_nexc)); + } + + + // Specializations for Heis states: + Vect Descendent_States_with_iK_Stepped_Up + (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc) + { + return(Descendent_States_with_iK_Stepped_Up (ScanIx2_label, OriginState.Ix2, OriginState.Ix2, OriginState.base.Ix2_min, OriginState.base.Ix2_max, disperse_only_current_exc, preserve_nexc)); + } + Vect Descendent_States_with_iK_Stepped_Down + (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc) + { + return(Descendent_States_with_iK_Stepped_Down (ScanIx2_label, OriginState.Ix2, OriginState.Ix2, OriginState.base.Ix2_min, OriginState.base.Ix2_max, disperse_only_current_exc, preserve_nexc)); + } + Vect Descendent_States_with_iK_Preserved + (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc_up, bool preserve_nexc_up, bool disperse_only_current_exc_down, bool preserve_nexc_down) + { + return(Descendent_States_with_iK_Preserved (ScanIx2_label, OriginState.Ix2, OriginState.Ix2, OriginState.base.Ix2_min, OriginState.base.Ix2_max, disperse_only_current_exc_up, preserve_nexc_up, disperse_only_current_exc_down, preserve_nexc_down)); + } + Vect Descendent_States_with_iK_Stepped_Up_rightIx2only + (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc) + { + return(Descendent_States_with_iK_Stepped_Up_rightIx2only (ScanIx2_label, OriginState.Ix2, OriginState.Ix2, OriginState.base.Ix2_min, OriginState.base.Ix2_max, disperse_only_current_exc, preserve_nexc)); + } + Vect Descendent_States_with_iK_Stepped_Down_rightIx2only + (string ScanIx2_label, const Heis_Bethe_State& OriginState, bool disperse_only_current_exc, bool preserve_nexc) + { + return(Descendent_States_with_iK_Stepped_Down_rightIx2only (ScanIx2_label, OriginState.Ix2, OriginState.Ix2, OriginState.base.Ix2_min, OriginState.base.Ix2_max, disperse_only_current_exc, preserve_nexc)); + } + + + + Vect Is_Good_New_Hole_Position (const Vect >& OriginIx2, State_Label_Data currentdata, int exclevel_newph) + { + // Given a state, returns the acceptable new hole positions. + + + // Define the objects for the newstatedata: + Vect type_new = currentdata.type; + Vect M_new = currentdata.M; + Vect nexc_new = currentdata.nexc; + nexc_new[exclevel_newph] += 1; // we drill one more particle-hole pair at this level + int index_new = (currentdata.nexc[exclevel_newph] + 1)/2; // we put the new p-h pair at index index_new. + + //int ntypespresent = ScanIx2.size(); + int ntypespresent = currentdata.type.size(); + Vect > Ix2old_new(ntypespresent); + Vect > Ix2exc_new(ntypespresent); + for (int it = 0; it < ntypespresent; ++it) Ix2old_new[it] = Vect(JSC::max(nexc_new[it],1)); + for (int it = 0; it < ntypespresent; ++it) Ix2exc_new[it] = Vect(JSC::max(nexc_new[it],1)); + + // Copy earlier data in: + for (int it = 0; it < ntypespresent; ++it) { + for (int i = 0; i < currentdata.nexc[it]; ++i) { + Ix2old_new[it][i + (it == exclevel_newph && i >= index_new)] = currentdata.Ix2old[it][i]; + Ix2exc_new[it][i + (it == exclevel_newph && i >= index_new)] = currentdata.Ix2exc[it][i]; + } + } + + //cout << "Here 1" << endl; + + State_Label_Data descdatanewph (type_new, M_new, nexc_new, Ix2old_new, Ix2exc_new); + + // We now look for all possible hole positions, + // the allowable ones being either at the edge of a block, or next to an existing hole excitation, + // under the condition of obeying the `towards the block center' rule. + + Vect isgoodnewholepos(false, OriginIx2[descdatanewph.type[exclevel_newph] ].size()); + + for (int ih = 0; ih < OriginIx2[descdatanewph.type[exclevel_newph] ].size(); ++ih) { + + int candidateIx2old = OriginIx2[descdatanewph.type[exclevel_newph] ][ih]; + + // candidateIx2old is an acceptable position for the new hole provided the following conditions are fulfilled: + // A- it is in OriginIx2 + // B- it follows the ordering rule, i.e. it sits in the middle of previous particle excitations, + // namely between Ix2old[index_new - 1] (if this exists) and Ix2old[index_new] (if this exists) + // C- it is next to an OriginIx2 vacancy or immediately right of Ix2old[index_new - 1] (if this exists) + // or immediately left of Ix2old[index_new] (if this exists) + // D- it does not break the `towards the center' rule + // (it will break the rule at this point if it is created away from an OriginIx2 boundary (and thus next to a preexisting excitation), + // and if this excitation and it are not in the same sideblock (in other words: if there is a sideblock boundary between them) + + //cout << "candidateIx2old " << candidateIx2old << " being tested" << endl; + // A + if (!OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old)) { + // is not contained in OriginIx2 + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason A" << endl; + continue; + } + + // B1 + if (currentdata.nexc[exclevel_newph] > 0 + && candidateIx2old <= currentdata.Ix2old[exclevel_newph][index_new - 1]) { + // there is at least one hole exc to the left, and the candidate position isn't right of Ix2old[index_new - 1] + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason B1" << endl; + continue; + } + + // B2 + if (currentdata.nexc[exclevel_newph] > 1 + && candidateIx2old >= currentdata.Ix2old[exclevel_newph][index_new]) { + // there is at least one hole exc to the right, and the candidate position isn't left of Ix2old[index_new] + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason B2" << endl; + continue; + } + + // C- it is next to an OriginIx2 vacancy or immediately right of Ix2old[index_new - 1] (if this exists) + // or immediately left of Ix2old[index_new] (if this exists) + if (!(!OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old + 2) || !OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old - 2)) + // doesn't sit next to an OriginIx2 vacancy + && (currentdata.nexc[exclevel_newph] == 0 || candidateIx2old != currentdata.Ix2old[exclevel_newph][index_new - 1] + 2) + // doesn't sit immediately right of first hole excitation to the left + && (currentdata.nexc[exclevel_newph] <= 1 || candidateIx2old != currentdata.Ix2old[exclevel_newph][index_new] - 2) + // doesn't sit immediately left of first hole excitation to the right + ) { + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason C" << endl; + continue; + } + + // D- it does not break the `towards the center' rule + // In other words, if created away from a block boundary but next to a preexisting hole, + // must be in same sideblock as this particle: + + // Determine the size of the block of OriginIx2 in which this hole sits: + int nroccupiedtoleft = 0; + while (OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old -2*(nroccupiedtoleft + 1))) + nroccupiedtoleft++; + int nroccupiedtoright = 0; + while (OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old +2*(nroccupiedtoright + 1))) + nroccupiedtoright++; + + // We can determine whether the new hole would be left- or right-moving + bool hole_candidate_is_left_moving = nroccupiedtoleft >= nroccupiedtoright; + bool hole_candidate_is_right_moving = nroccupiedtoleft < nroccupiedtoright; + + if (hole_candidate_is_left_moving + && (currentdata.nexc[exclevel_newph] > 0 && candidateIx2old == currentdata.Ix2old[exclevel_newph][index_new - 1] + 2) + // it is created to the right of a preexisting hole excitation + && OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old + 2) // and is not sitting at the boundary + && (currentdata.nexc[exclevel_newph] <= 1 || candidateIx2old != currentdata.Ix2old[exclevel_newph][index_new] - 2) + // and is not sitting just left of another preexisting hole excitation which is also left moving + ) + { + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason D1" << endl; + continue; + } + + if (hole_candidate_is_right_moving + && (currentdata.nexc[exclevel_newph] > 1 && candidateIx2old == currentdata.Ix2old[exclevel_newph][index_new] - 2) + // it is created to the left of a preexisting hole excitation + && OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old - 2) // and is not sitting at the boundary + && (currentdata.nexc[exclevel_newph] == 0 || candidateIx2old != currentdata.Ix2old[exclevel_newph][index_new - 1] + 2) + // and is not sitting just right of another preexisting hole excitation which is also right moving + ) + { + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason D2" << endl; + continue; + } + + + // If we have reached this point, candidateIx2old is acceptable. + + isgoodnewholepos[ih] = true; + + } // for (int ih + + return(isgoodnewholepos); + + } // Vect Is_Good_New_Hole_Position + + + + + //Vect Descendents (const Bethe_State& ScanState, string type_required) + Vect Descendents (string ScanStateLabel, const Vect >& ScanIx2, const Vect >& OriginIx2, + const Vect& Ix2_min, const Vect& Ix2_max, int type_required) + { + // This generates descendents for a fixed base. + // As compared to the OriginState, ScanState thus has only (in-level) particle-hole excitations + + // ASSUMPTIONS: + // ScanStateLabel is consistent with ScanIx2 and OriginIx2. + + // type_required == 0: inner (right)most hole moved + // type_required == 1: inner (right)most particle moved + // type_required == 2: new particle-hole pair added, keeping only closest p < h and p > h cases + // type_required == 3: at level 0, move leftmost particle one step left, rightmost one step right (fixed iK logic using skeletons). + // type_required == 4: generalized Umklapp: increase distance between latest-added p-h pair, staying on boundary of blocks + + //cout << "\tCalling Descendents for type = " << type_required << endl; + + bool cinbreaks = false; + + // Number of descendents: + //int ndesc_possible = (type_required == 0 ? 1 : ScanState.base.charge * 2); // upper bound guess; should be refined. + //int ndesc_possible = (type_required == 0 ? 1 : 2* ScanIx2[0].size() * 2 * ScanIx2[0].size()); // upper bound guess; should be refined + int ndesc_possible = 1; + if (type_required == 1 || type_required == 2) { + for (int i = 0; i < ScanIx2.size(); ++i) + ndesc_possible = JSC::max (ndesc_possible, 2* ScanIx2[i].size() * 2 * ScanIx2[i].size()); // inexact, should be refined + } + //cout << "ndesc_possible = " << ndesc_possible << endl; + + //cout << "OriginIx2.size() = " << OriginIx2.size() << endl; + //for (int i = 0; i < OriginIx2.size(); ++i) cout << "i = " << i << "\tOriginIx2[i].size() = " << OriginIx2[i].size() << endl; + //for (int i = 0; i < OriginIx2.size(); ++i) cout << "i = " << i << "\tOriginIx2[i] = " << OriginIx2[i] << endl; + //cout << "OriginIx2 = " << endl; + //cout << OriginIx2 << endl; + + Vect desclabelfound (ndesc_possible); + Vect desctypefound (ndesc_possible); + int ndesc_found = 0; + + // First read the current state data: + State_Label_Data currentdata = Read_State_Label (ScanStateLabel, OriginIx2); + + // Determine the level at which the highest current excitation sits + //int exclevel = ScanIx2.size() - 1; + int exclevel = currentdata.type.size() - 1; + while (exclevel > 0 && currentdata.nexc[exclevel] == 0) exclevel--; + //cout << "exclevel = " << exclevel << endl; + + //cout << "***** Looking for descendents of type " << type_required << " for state with OriginIx2, ScanIx2: *******" << endl; + //cout << OriginIx2[currentdata.type[exclevel] ] << endl; + //cout << ScanIx2[currentdata.type[exclevel] ] << endl; + //if (currentdata.nexc[exclevel] > 0) { + //cout << "with current excitations (holes, particles):" << endl; + //cout << currentdata.Ix2old[exclevel] << endl; + //cout << currentdata.Ix2exc[exclevel] << endl; + //} + + if (type_required == 0) { + + // We move the inner(right)most hole of the highest excited level one more step towards + // the center of the block of Ix2 vacancies in which it sits. + + // Is there already an excitation at level 0? If so, move inner(right)most particle one step further. + if (currentdata.nexc[0] > 0) { + + // Produce a new state_label_data: + State_Label_Data descdata = Read_State_Label (ScanStateLabel, OriginIx2); + // Identify the inner(right)most excitation: + int innerindex = currentdata.nexc[exclevel]/2; + + int newholepos = currentdata.Ix2old[exclevel][innerindex]; + + // Determine the size of the block of OriginState.Ix2 in which this hole sits: + int nroccupiedtoleft = 0; + while (OriginIx2[descdata.type[exclevel] ].includes (currentdata.Ix2old[exclevel][innerindex] -2*(nroccupiedtoleft + 1))) + nroccupiedtoleft++; + int nroccupiedtoright = 0; + while (OriginIx2[descdata.type[exclevel] ].includes (currentdata.Ix2old[exclevel][innerindex] +2*(nroccupiedtoright + 1))) + nroccupiedtoright++; + + if (nroccupiedtoleft - nroccupiedtoright + 2 < 0) newholepos += 2; + else if (nroccupiedtoleft - nroccupiedtoright - 2 >= 0) newholepos -= 2; + + if (newholepos != currentdata.Ix2old[exclevel][innerindex] // we have successfully moved the hole + && !currentdata.Ix2old[exclevel].includes(newholepos) // new hole position is not already taken + ) + { // we have found a descendent + descdata.Ix2old[exclevel][innerindex] = newholepos; + desclabelfound[ndesc_found] = Return_State_Label (descdata, OriginIx2); + desctypefound[ndesc_found] = 0; + //cout << "For state with OriginIx2, ScanIx2:" << endl; + //cout << OriginIx2[exclevel] << endl; + //cout << ScanIx2[exclevel] << endl; + //if (currentdata.nexc[exclevel] > 0) { + //cout << "with current excitations (holes, particles):" << endl; + //cout << currentdata.Ix2old[exclevel] << endl; + //cout << currentdata.Ix2exc[exclevel] << endl; + //} + //cout << "\tFound new descendent of type " << desctypefound[ndesc_found] << ": label " << desclabelfound[ndesc_found] << "\tholes " << descdata.Ix2old[exclevel] << "\tpart " << descdata.Ix2exc[exclevel] << endl; + ndesc_found++; + if (cinbreaks) { char a; cin >> a;} + } + + } // if (currentdata.nexc[exclevel] > 0) + + } // if (type_required == 0) + + + if (type_required == 1) { + + //cout << "exclevel = " << exclevel << endl; + //cout << "currentdata.nexc[exclevel] = " << currentdata.nexc[exclevel] << endl; + + // We move the inner(right)most particle of the highest excited level one more step towards + // the center of the block of Ix2 vacancies in which it sits. + + // Is there already an excitation at level 0? If so, move inner(right)most particle one step further. + if (currentdata.nexc[exclevel] > 0) { + + // Produce a new state_label_data: + State_Label_Data descdata = Read_State_Label (ScanStateLabel, OriginIx2); + // Identify the inner(right)most excitation: + int innerindex = currentdata.nexc[exclevel]/2; + + int partpos = currentdata.Ix2exc[exclevel][innerindex]; + + // Determine the size of the block of OriginState.Ix2 vacancies in which this particle sits: + bool partpos_is_left_of_all_OriginIx2 = (partpos <= OriginIx2[descdata.type[exclevel] ].min()); + bool partpos_is_right_of_all_OriginIx2 = (partpos >= OriginIx2[descdata.type[exclevel] ].max()); + int nremptytoleft = 0; + if (partpos_is_left_of_all_OriginIx2) nremptytoleft = (partpos - Ix2_min[descdata.type[exclevel] ])/2; + else while (!OriginIx2[descdata.type[exclevel] ].includes (partpos -2*(nremptytoleft + 1))) + nremptytoleft++; + int nremptytoright = 0; + if (partpos_is_right_of_all_OriginIx2) nremptytoright = (Ix2_max[descdata.type[exclevel] ] - partpos)/2; + else while (!OriginIx2[descdata.type[exclevel] ].includes (partpos +2*(nremptytoright + 1))) + nremptytoright++; + + if (!partpos_is_left_of_all_OriginIx2 && (partpos_is_right_of_all_OriginIx2 || nremptytoleft - nremptytoright + 2 < 0)) partpos += 2; + else if (!partpos_is_right_of_all_OriginIx2 && (partpos_is_left_of_all_OriginIx2 || nremptytoleft - nremptytoright - 2 >= 0)) partpos -= 2; + + if (partpos != currentdata.Ix2exc[exclevel][innerindex] // we have successfully moved the particle + && !OriginIx2[descdata.type[exclevel] ].includes(partpos) // it's actually a new particle position + && !currentdata.Ix2exc[exclevel].includes(partpos) // new particle position is not already taken + && partpos >= Ix2_min[descdata.type[exclevel] ] + && partpos <= Ix2_max[descdata.type[exclevel] ] + ) + { // we have found a descendent + descdata.Ix2exc[exclevel][innerindex] = partpos; + desclabelfound[ndesc_found] = Return_State_Label (descdata, OriginIx2); + desctypefound[ndesc_found] = 1; + //cout << "For state with OriginIx2, ScanIx2:" << endl; + //cout << OriginIx2[exclevel] << endl; + //cout << ScanIx2[exclevel] << endl; + //if (currentdata.nexc[exclevel] > 0) { + //cout << "with current excitations (holes, particles):" << endl; + //cout << currentdata.Ix2old[exclevel] << endl; + //cout << currentdata.Ix2exc[exclevel] << endl; + //} + //cout << "\tFound new descendent of type " << desctypefound[ndesc_found] << ": label " << desclabelfound[ndesc_found] << "\tholes " << descdata.Ix2old[exclevel] << "\tpart " << descdata.Ix2exc[exclevel] << endl; + ndesc_found++; + if (cinbreaks) { char a; cin >> a;} + } + + } // if (currentdata.nexc[exclevel] > 0) + + + } // if (type_required == 1) + + + if (type_required == 2) { + + // Now add a new p-h pair at the inner(right)most position, at each level from exclevel upwards, + // putting the particle and hole to each available positions at the edge of vacancy/occupancy blocks. + + //cout << "Trying for type 2 descendent. exclevel = " << exclevel << "\tcurrentdata.nexc.size() = " << currentdata.nexc.size() << endl; + + //for (int exclevel_newph = exclevel; exclevel_newph < ScanIx2.size(); ++exclevel_newph) { + for (int exclevel_newph = exclevel; exclevel_newph < currentdata.nexc.size(); ++exclevel_newph) { + + if (ScanIx2[currentdata.type[exclevel_newph] ].size() <= currentdata.nexc[exclevel_newph]) continue; // no space for another p-h + + // Define the objects for the newstatedata: + Vect type_new = currentdata.type; + Vect M_new = currentdata.M; + Vect nexc_new = currentdata.nexc; + nexc_new[exclevel_newph] += 1; // we drill one more particle-hole pair at this level + int index_new = (currentdata.nexc[exclevel_newph] + 1)/2; // we put the new p-h pair at index index_new. + + //int ntypespresent = ScanIx2.size(); + int ntypespresent = currentdata.type.size(); + Vect > Ix2old_new(ntypespresent); + Vect > Ix2exc_new(ntypespresent); + for (int it = 0; it < ntypespresent; ++it) Ix2old_new[it] = Vect(JSC::max(nexc_new[it],1)); + for (int it = 0; it < ntypespresent; ++it) Ix2exc_new[it] = Vect(JSC::max(nexc_new[it],1)); + + // Copy earlier data in: + for (int it = 0; it < ntypespresent; ++it) { + for (int i = 0; i < currentdata.nexc[it]; ++i) { + Ix2old_new[it][i + (it == exclevel_newph && i >= index_new)] = currentdata.Ix2old[it][i]; + Ix2exc_new[it][i + (it == exclevel_newph && i >= index_new)] = currentdata.Ix2exc[it][i]; + } + } + + //cout << "Here 1" << endl; + + State_Label_Data descdatanewph (type_new, M_new, nexc_new, Ix2old_new, Ix2exc_new); + + // We now look for all possible hole positions, + // the allowable ones being either at the edge of a block, or next to an existing hole excitation, + // under the condition of obeying the `towards the block center' rule. + + Vect isgoodnewholepos = Is_Good_New_Hole_Position (OriginIx2, currentdata, exclevel_newph); + /* The function call above replaces the whole block here: + Vect isgoodnewholepos(false, OriginIx2[descdatanewph.type[exclevel_newph] ].size()); + + for (int ih = 0; ih < OriginIx2[descdatanewph.type[exclevel_newph] ].size(); ++ih) { + + int candidateIx2old = OriginIx2[descdatanewph.type[exclevel_newph] ][ih]; + + // candidateIx2old is an acceptable position for the new hole provided the following conditions are fulfilled: + // A- it is in OriginIx2 + // B- it follows the ordering rule, i.e. it sits in the middle of previous particle excitations, + // namely between Ix2old[index_new - 1] (if this exists) and Ix2old[index_new] (if this exists) + // C- it is next to an OriginIx2 vacancy or immediately right of Ix2old[index_new - 1] (if this exists) + // or immediately left of Ix2old[index_new] (if this exists) + // D- it does not break the `towards the center' rule + // (it will break the rule at this point if it is created away from an OriginIx2 boundary (and thus next to a preexisting excitation), + // and if this excitation and it are not in the same sideblock (in other words: if there is a sideblock boundary between them) + + //cout << "candidateIx2old " << candidateIx2old << " being tested" << endl; + // A + if (!OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old)) { + // is not contained in OriginIx2 + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason A" << endl; + continue; + } + + // B1 + if (currentdata.nexc[exclevel_newph] > 0 + && candidateIx2old <= currentdata.Ix2old[exclevel_newph][index_new - 1]) { + // there is at least one hole exc to the left, and the candidate position isn't right of Ix2old[index_new - 1] + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason B1" << endl; + continue; + } + + // B2 + if (currentdata.nexc[exclevel_newph] > 1 + && candidateIx2old >= currentdata.Ix2old[exclevel_newph][index_new]) { + // there is at least one hole exc to the right, and the candidate position isn't left of Ix2old[index_new] + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason B2" << endl; + continue; + } + + // C- it is next to an OriginIx2 vacancy or immediately right of Ix2old[index_new - 1] (if this exists) + // or immediately left of Ix2old[index_new] (if this exists) + if (!(!OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old + 2) || !OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old - 2)) + // doesn't sit next to an OriginIx2 vacancy + && (currentdata.nexc[exclevel_newph] == 0 || candidateIx2old != currentdata.Ix2old[exclevel_newph][index_new - 1] + 2) + // doesn't sit immediately right of first hole excitation to the left + && (currentdata.nexc[exclevel_newph] <= 1 || candidateIx2old != currentdata.Ix2old[exclevel_newph][index_new] - 2) + // doesn't sit immediately left of first hole excitation to the right + ) { + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason C" << endl; + continue; + } + + // D- it does not break the `towards the center' rule + // In other words, if created away from a block boundary but next to a preexisting hole, + // must be in same sideblock as this particle: + + // Determine the size of the block of OriginIx2 in which this hole sits: + int nroccupiedtoleft = 0; + while (OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old -2*(nroccupiedtoleft + 1))) + nroccupiedtoleft++; + int nroccupiedtoright = 0; + while (OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old +2*(nroccupiedtoright + 1))) + nroccupiedtoright++; + + // We can determine whether the new hole would be left- or right-moving + bool hole_candidate_is_left_moving = nroccupiedtoleft >= nroccupiedtoright; + bool hole_candidate_is_right_moving = nroccupiedtoleft < nroccupiedtoright; + + if (hole_candidate_is_left_moving + && (currentdata.nexc[exclevel_newph] > 0 && candidateIx2old == currentdata.Ix2old[exclevel_newph][index_new - 1] + 2) + // it is created to the right of a preexisting hole excitation + && OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old + 2) // and is not sitting at the boundary + && (currentdata.nexc[exclevel_newph] <= 1 || candidateIx2old != currentdata.Ix2old[exclevel_newph][index_new] - 2) + // and is not sitting just left of another preexisting hole excitation which is also left moving + ) + { + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason D1" << endl; + continue; + } + + if (hole_candidate_is_right_moving + && (currentdata.nexc[exclevel_newph] > 1 && candidateIx2old == currentdata.Ix2old[exclevel_newph][index_new] - 2) + // it is created to the left of a preexisting hole excitation + && OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2old - 2) // and is not sitting at the boundary + && (currentdata.nexc[exclevel_newph] == 0 || candidateIx2old != currentdata.Ix2old[exclevel_newph][index_new - 1] + 2) + // and is not sitting just right of another preexisting hole excitation which is also right moving + ) + { + //cout << "candidateIx2old " << candidateIx2old << " rejected for reason D2" << endl; + continue; + } + + + // If we have reached this point, candidateIx2old is acceptable. + + isgoodnewholepos[ih] = true; + + } // for (int ih + */ + + + + //cout << "Here 2" << endl; + + //cout << "\tisgoodnewholdpos = " << isgoodnewholepos << endl; + + //cout << "Looking for a new particle position: OriginIx2, ScanIx2:" << endl; + //cout << OriginIx2[exclevel_newph] << endl; + //cout << ScanIx2[exclevel_newph] << endl; + //if (currentdata.nexc[exclevel_newph] > 0) { + // cout << "Current excitations (holes, particles):" << endl; + // cout << currentdata.Ix2old[exclevel_newph] << endl; + // cout << currentdata.Ix2exc[exclevel_newph] << endl; + //} + + // We now look for all possible particle positions, + // the allowable ones being either at the edge of a block, or next to an existing particle excitation, + // under the condition of obeying the `towards the block center' rule. + + // Determine range of possible Ix2exc: + int Ix2excmin = ScanIx2[descdatanewph.type[exclevel_newph] ].min() - 2; + if (currentdata.nexc[exclevel_newph] > 0) + Ix2excmin = JSC::min (Ix2excmin, currentdata.Ix2exc[exclevel_newph][index_new - 1] + 2); + Ix2excmin = JSC::max (Ix2excmin, Ix2_min[descdatanewph.type[exclevel_newph] ]); + int Ix2excmax = ScanIx2[descdatanewph.type[exclevel_newph] ].max() + 2; + if (currentdata.nexc[exclevel_newph] > 1) + Ix2excmax = JSC::max (Ix2excmax, currentdata.Ix2exc[exclevel_newph][index_new] - 2); + Ix2excmax = JSC::min (Ix2excmax, Ix2_max[descdatanewph.type[exclevel_newph] ]); + + Vect isgoodnewpartpos(false, (Ix2excmax - Ix2excmin)/2 + 1); + + for (int candidateIx2exc = Ix2excmin; candidateIx2exc <= Ix2excmax; candidateIx2exc += 2) { + + //cout << "Here a" << endl; + + // candidateIx2exc is an acceptable position for the new particle provided the following conditions are fulfilled: + // A- it is not in OriginIx2 + // B- it follows the ordering rule, i.e. it sits in the middle of previous particle excitations, + // namely between Ix2exc[index_new - 1] (if this exists) and Ix2exc[index_new] (if this exists) + // C- it is next to an OriginIx2 or immediately right of Ix2exc[index_new - 1] (if this exists) + // or immediately left of Ix2exc[index_new] (if this exists) + // D- it does not break the `towards the center' rule + // (it will break the rule at this point if it is created away from an OriginIx2 boundary (and thus next to a preexisting excitation), + // and if this excitation and it are not in the same sideblock (in other words: if there is a sideblock boundary between them) + + //cout << "candidateIx2exc " << candidateIx2exc << " being tested" << endl; + // A + if (OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2exc)) { + // is contained in OriginIx2 + //cout << "candidateIx2exc " << candidateIx2exc << " rejected for reason A" << endl; + continue; + } + + //cout << "Here b" << endl; + + // B1 + if (currentdata.nexc[exclevel_newph] > 0 + && candidateIx2exc <= currentdata.Ix2exc[exclevel_newph][index_new - 1]) { + // there is at least one particle exc to the left, and the candidate position isn't right of Ix2exc[index_new - 1] + //cout << "candidateIx2exc " << candidateIx2exc << " rejected for reason B1" << endl; + continue; + } + + //cout << "Here c" << endl; + + // B2 + if (currentdata.nexc[exclevel_newph] > 1 + && candidateIx2exc >= currentdata.Ix2exc[exclevel_newph][index_new]) { + // there is at least one particle exc to the right, and the candidate position isn't left of Ix2exc[index_new] + //cout << "candidateIx2exc " << candidateIx2exc << " rejected for reason B2" << endl; + continue; + } + + //cout << "Here d" << endl; + + // C- it is next to an OriginIx2 or immediately right of Ix2exc[index_new - 1] (if this exists) + // or immediately left of Ix2exc[index_new] (if this exists) + if (!(OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2exc + 2) || OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2exc - 2)) + // doesn't sit next to an OriginIx2 + && (currentdata.nexc[exclevel_newph] == 0 || candidateIx2exc != currentdata.Ix2exc[exclevel_newph][index_new - 1] + 2) + // doesn't sit immediately right of first particle excitation to the left + && (currentdata.nexc[exclevel_newph] <= 1 || candidateIx2exc != currentdata.Ix2exc[exclevel_newph][index_new] - 2) + // doesn't sit immediately left of first particle excitation to the right + ) { + //cout << "candidateIx2exc " << candidateIx2exc << " rejected for reason C" << endl; + continue; + } + + //cout << "Here e" << endl; + + // D- it does not break the `towards the center' rule + // In other words, if created away from a block boundary but next to a preexisting particle, + // must be in same sideblock as this particle: + + // Determine the size of the block of OriginIx2 vacancies in which this particle sits: + bool candidate_is_left_of_all_OriginIx2 = (candidateIx2exc < OriginIx2[descdatanewph.type[exclevel_newph] ].min()); // this makes it left-moving + bool candidate_is_right_of_all_OriginIx2 = (candidateIx2exc > OriginIx2[descdatanewph.type[exclevel_newph] ].max()); // this makes it right-moving + int nremptytoleft = 0; + if (candidate_is_left_of_all_OriginIx2) nremptytoleft = (candidateIx2exc - Ix2_min[descdatanewph.type[exclevel_newph] ])/2; + else while (!OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2exc -2*(nremptytoleft + 1))) + nremptytoleft++; + int nremptytoright = 0; + if (candidate_is_right_of_all_OriginIx2) nremptytoright = (Ix2_max[descdatanewph.type[exclevel_newph] ] - candidateIx2exc)/2; + else while (!OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2exc +2*(nremptytoright + 1))) + nremptytoright++; + // We can determine whether the new particle would be left- or right-moving + bool candidate_is_left_moving = candidate_is_left_of_all_OriginIx2 + || (!candidate_is_right_of_all_OriginIx2 && nremptytoleft >= nremptytoright); + bool candidate_is_right_moving = candidate_is_right_of_all_OriginIx2 + || (!candidate_is_left_of_all_OriginIx2 && nremptytoleft < nremptytoright); + // Consistency checks: + if (candidate_is_left_moving && candidate_is_right_moving) JSCerror("New particle moving left and right at same time"); + if (!candidate_is_left_moving && !candidate_is_right_moving) JSCerror("New particle not moving either left or right"); + + //cout << "Here f" << endl; + + if (candidate_is_left_moving + && (currentdata.nexc[exclevel_newph] > 0 && candidateIx2exc == currentdata.Ix2exc[exclevel_newph][index_new - 1] + 2) + // it is created to the right of a preexisting particle excitation + && !OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2exc + 2) // and is not sitting at the boundary + && (currentdata.nexc[exclevel_newph] <= 1 || candidateIx2exc != currentdata.Ix2exc[exclevel_newph][index_new] - 2) + // and is not sitting just left of another preexisting particle excitation which is also left moving + ) + { + //cout << "candidateIx2exc " << candidateIx2exc << " rejected for reason D1" << endl; + continue; + } + + //cout << "Here g" << endl; + + if (candidate_is_right_moving + && (currentdata.nexc[exclevel_newph] > 1 && candidateIx2exc == currentdata.Ix2exc[exclevel_newph][index_new] - 2) + // it is created to the left of a preexisting particle excitation + && !OriginIx2[descdatanewph.type[exclevel_newph] ].includes (candidateIx2exc - 2) // and is not sitting at the boundary + && (currentdata.nexc[exclevel_newph] == 0 || candidateIx2exc != currentdata.Ix2exc[exclevel_newph][index_new - 1] + 2) + // and is not sitting just right of another preexisting particle excitation which is also right moving + ) + { + //cout << "candidateIx2exc " << candidateIx2exc << " rejected for reason D2" << endl; + continue; + } + + //cout << "\tFound a possible exc position at " << candidateIx2exc << endl; + + // If we have reached this point, candidateIx2exc is acceptable. + // Immediately construct all descendents with this part position: + + //cout << "Here h" << endl; + + // We now select the hole position closest to left or right: + int ihclosestleft = -1; + int ihclosestright = -1; + for (int ih = 0; ih < OriginIx2[descdatanewph.type[exclevel_newph] ].size(); ++ih) + if (isgoodnewholepos[ih]) + { + /* + //cout << "Here alpha" << endl; + descdatanewph.Ix2old[exclevel_newph][index_new] = OriginIx2[descdatanewph.type[exclevel_newph] ][ih]; + //cout << "Here beta" << endl; + descdatanewph.Ix2exc[exclevel_newph][index_new] = candidateIx2exc; + //cout << "Here gamma" << endl; + desclabelfound[ndesc_found] = Return_State_Label (descdatanewph, OriginIx2); + //cout << "Here delta" << endl; + desctypefound[ndesc_found] = 2; + //cout << "For state with OriginIx2, ScanIx2:" << endl; + //cout << OriginIx2[exclevel_newph] << endl; + //cout << ScanIx2[exclevel_newph] << endl; + //if (currentdata.nexc[exclevel_newph] > 0) { + // cout << "with current excitations (holes, particles):" << endl; + // cout << currentdata.Ix2old[exclevel_newph] << endl; + // cout << currentdata.Ix2exc[exclevel_newph] << endl; + //} + //cout << "\tFound new descendent of type " << desctypefound[ndesc_found] << ": label " << desclabelfound[ndesc_found] << "\tholes " << descdatanewph.Ix2old[exclevel_newph] << "\tpart " << descdatanewph.Ix2exc[exclevel_newph] << endl; + ndesc_found++; + + if (cinbreaks) { char a; cin >> a;} + */ + // New in ++G_3.1: + if (OriginIx2[descdatanewph.type[exclevel_newph] ][ih] - candidateIx2exc < 0) // new hole is left of new particle + if (ihclosestleft == -1 // we hadn't found a hole previously + || ihclosestleft >= 0 // we had aleady found a new hole to the left + && OriginIx2[descdatanewph.type[exclevel_newph] ][ih] > OriginIx2[descdatanewph.type[exclevel_newph] ][ihclosestleft]) // + ihclosestleft = ih; + if (OriginIx2[descdatanewph.type[exclevel_newph] ][ih] - candidateIx2exc > 0) // new hole is right of new particle + if (ihclosestright == -1 // we hadn't found a hole previously + || ihclosestright >= 0 // we had aleady found a new hole to the right + && OriginIx2[descdatanewph.type[exclevel_newph] ][ih] < OriginIx2[descdatanewph.type[exclevel_newph] ][ihclosestright]) // + ihclosestright = ih; + } // if (isgoodnewholepos[ih]) + // for ih + + if (ihclosestleft >= 0) { // Found a descendent with new hole to left of new particle + descdatanewph.Ix2old[exclevel_newph][index_new] = OriginIx2[descdatanewph.type[exclevel_newph] ][ihclosestleft]; + descdatanewph.Ix2exc[exclevel_newph][index_new] = candidateIx2exc; + desclabelfound[ndesc_found] = Return_State_Label (descdatanewph, OriginIx2); + desctypefound[ndesc_found] = 2; + ndesc_found++; + } + + if (ihclosestright >= 0) { // Found a descendent with new hole to right of new particle + descdatanewph.Ix2old[exclevel_newph][index_new] = OriginIx2[descdatanewph.type[exclevel_newph] ][ihclosestright]; + descdatanewph.Ix2exc[exclevel_newph][index_new] = candidateIx2exc; + desclabelfound[ndesc_found] = Return_State_Label (descdatanewph, OriginIx2); + desctypefound[ndesc_found] = 2; + ndesc_found++; + } + + } // for (int candidateIx2exc) + + //cout << "Here 3" << endl; + + } // for (int exclevel_newph) + + } // if (type_required == 2) + + + if (type_required == 3) { + + // We here only move the leftmost Ix2[0] left, and rightmost Ix2[0] right (this preserves iK) + + if (ScanIx2[0][0] > Ix2_min[0] && ScanIx2[0][ScanIx2[0].size() - 1] < Ix2_max[0]) { + Vect > descIx2 = ScanIx2; + descIx2[0][0] -= 2; + descIx2[0][descIx2[0].size() - 1] += 2; + desclabelfound[ndesc_found] = Return_State_Label (descIx2, OriginIx2); + desctypefound[ndesc_found] = 3; + ndesc_found++; + } + + } // if (type_required == 3) + + + if (type_required == 4) { + + // We take the latest particle-hole pair, and move the hole to a further generalized Umklapp position + + State_Label_Data descdata = currentdata; + + // Start by finding the possible hole positions of the state without the innermost p-h pair: + Vect nexc_new = currentdata.nexc; + if (currentdata.nexc[exclevel] < 1) JSCerror("Should not call type 3 descendent if there are no ph exc at exclevel"); + nexc_new[exclevel] -= 1; // we remove the innermost ph pair + + int innerindex = currentdata.nexc[exclevel]/2; // index of ph pair we remove + + int ntypespresent = currentdata.type.size(); + Vect > Ix2old_new(ntypespresent); + Vect > Ix2exc_new(ntypespresent); + for (int it = 0; it < ntypespresent; ++it) Ix2old_new[it] = Vect(JSC::max(nexc_new[it],1)); + for (int it = 0; it < ntypespresent; ++it) Ix2exc_new[it] = Vect(JSC::max(nexc_new[it],1)); + + // Copy earlier data in: + for (int it = 0; it < ntypespresent; ++it) { + for (int i = 0; i < nexc_new[it]; ++i) { + Ix2old_new[it][i] = currentdata.Ix2old[it][i + (it == exclevel && i >= innerindex)]; + Ix2exc_new[it][i] = currentdata.Ix2exc[it][i + (it == exclevel && i >= innerindex)]; + } + } + + State_Label_Data data_latest_ph_removed (currentdata.type, currentdata.M, nexc_new, Ix2old_new, Ix2exc_new); + + Vect isgoodnewholepos = Is_Good_New_Hole_Position (OriginIx2, data_latest_ph_removed, exclevel); + + // Move hole further away from its corresponding particle, preserving the direction + bool holemovingleft = currentdata.Ix2old[exclevel][innerindex] < currentdata.Ix2exc[exclevel][innerindex]; + + int inewhole; + if (holemovingleft) { + inewhole = 0; + while (OriginIx2[currentdata.type[exclevel] ][inewhole] != currentdata.Ix2old[exclevel][innerindex]) inewhole++; + do { + inewhole--; + } while (inewhole >= 0 && !isgoodnewholepos[inewhole]); + } + + if (!holemovingleft) { + inewhole = OriginIx2[currentdata.type[exclevel] ].size() - 1; + while (OriginIx2[currentdata.type[exclevel] ][inewhole] != currentdata.Ix2old[exclevel][innerindex]) inewhole--; + do { + inewhole++; + } while (inewhole <= OriginIx2[currentdata.type[exclevel] ].size() - 1 && !isgoodnewholepos[inewhole]); + } + + if (inewhole >= 0 && inewhole <= OriginIx2[currentdata.type[exclevel] ].size() - 1) { // Found a descendent + descdata.Ix2old[exclevel][innerindex] = OriginIx2[currentdata.type[exclevel] ][inewhole]; + desclabelfound[ndesc_found] = Return_State_Label (descdata, OriginIx2); + desctypefound[ndesc_found] = 4; + ndesc_found++; + } + + } // if (type_required == 4) + + + + + Vect desclabelfoundresized (ndesc_found); + for (int i = 0; i < ndesc_found; ++i) { + desclabelfoundresized[i] = desclabelfound[i]; + } + + //cout << "For state with label " << ScanStateLabel << ", found " << ndesc_found << " descendents of type " << type_required << endl; + //cout << desclabelfoundresized << endl; + if (cinbreaks) { char a; cin >> a;} + + return(desclabelfoundresized); + + } // Vect Descendents + + + + // Specialization for LiebLin gas: + + Vect Descendents (const LiebLin_Bethe_State& ScanState, const LiebLin_Bethe_State& OriginState, int type_required) + //string ScanStateLabel, const Vect >& ScanIx2, const Vect >& OriginIx2, + // const Vect& Ix2_min, const Vect& Ix2_max, int type_required) + { + Vect > ScanIx2here(1); + ScanIx2here[0] = ScanState.Ix2; + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginState.Ix2; + Vect Ix2_min(1); + Ix2_min[0] = LIEBLIN_Ix2_MIN; + Vect Ix2_max(1); + Ix2_max[0] = LIEBLIN_Ix2_MAX; + Vect desc_found; + + // If the state is an outer skeleton, we use a slightly modified OriginIx2 (in which the outermost + // rapidities at level zero are put to the outer skeleton position convention) for computing the descendents. + + //if (ScanState.Is_Outer_Skeleton()) { // since the state is an outer skeleton, we first modify the OriginIx2 + //if (ScanState.N >= 2 && ScanState.Ix2[0] == LIEBLIN_Ix2_MIN + (ScanState.N % 2) + 1 && ScanState.Ix2[ScanState.N-1] == LIEBLIN_Ix2_MAX - (ScanState.N % 2) - 1) { // since the state is an outer skeleton, we first modify the OriginIx2 + if (Is_Outer_Skeleton(ScanState)) { + //cout << "\tDetected outer skeleton in Descendents: " << ScanState.label << endl; + /* First (failed) attempt + OriginIx2here[0][0] = LIEBLIN_Ix2_MIN + (ScanState.N % 2) + 3; // The OriginIx2here endpoints are set one step internally next to the Skeleton position conventions + OriginIx2here[0][OriginState.N - 1] = LIEBLIN_Ix2_MAX - (ScanState.N % 2) - 3; + // Obtain the descendends using this modified OriginIx2: + Vect desc_mod = Descendents (ScanState.label, ScanIx2here, OriginIx2here, Ix2_min, Ix2_max, type_required); + // Translate the obtained descendends into ones corresponding to OriginIx2: + LiebLin_Bethe_State LabellingState = ScanState; + for (int idesc = 0; idesc < desc_mod.size(); ++idesc) { + LabellingState.Set_to_Label (desc_mod[idesc], OriginIx2here[0]); // Set quantum numbers according to OriginIx2here + LabellingState.Set_Label_from_Ix2 (OriginState.Ix2); // Now reset label on OriginIx2 from the correctly set Ix2 + desc_mod[idesc] = LabellingState.label; + cout << "\tidesc " << idesc << "\tlabel " << LabellingState.label << "\tIx2 " << LabellingState.Ix2 << endl; + } + return(desc_mod); + */ + // Construct and descend a state with 2 less particles: + if (ScanState.N < 3) JSCerror("Skeleton descendent logic at fixed iK not implemented for N < 3."); + LiebLin_Bethe_State ReducedScanState (ScanState.c_int, ScanState.L, ScanState.N - 2); + LiebLin_Bethe_State ReducedOriginState (ScanState.c_int, ScanState.L, ScanState.N - 2); + for (int i = 0; i < ScanState.N - 2; ++i) { + ReducedScanState.Ix2[i] = ScanState.Ix2[i+1]; + ReducedOriginState.Ix2[i] = OriginState.Ix2[i+1]; + } + ReducedScanState.Set_Label_from_Ix2 (ReducedOriginState.Ix2); + Vect desc_mod = Descendents (ReducedScanState, ReducedOriginState, type_required); + // Now translate results back to ScanState: we return outer skeleton states + LiebLin_Bethe_State LabellingState = ScanState; + for (int idesc = 0; idesc < desc_mod.size(); ++idesc) { + ReducedScanState.Set_to_Label (desc_mod[idesc], ReducedOriginState.Ix2); + // Now set all quantum numbers of LabellingState: + LabellingState.Ix2[0] = LIEBLIN_Ix2_MIN + (ScanState.N % 2) + 1; + for (int i = 0; i < ScanState.N - 2; ++i) LabellingState.Ix2[i+1] = ReducedScanState.Ix2[i]; + LabellingState.Ix2[ScanState.N - 1] = LIEBLIN_Ix2_MAX - (ScanState.N % 2) - 1; + LabellingState.Set_Label_from_Ix2 (OriginState.Ix2); // Now reset label on OriginIx2 from the correctly set Ix2 + desc_mod[idesc] = LabellingState.label; + //cout << "\tidesc " << idesc << "\tlabel " << LabellingState.label << "\tIx2 " << LabellingState.Ix2 << endl; + } + //cout << "\tFound " << desc_mod.size() << "descendents, " << endl; + //cout << "\t" << desc_mod << endl; + desc_found = desc_mod; + } + + // If not outer skeleton, just return straight descendents + else desc_found = Descendents (ScanState.label, ScanIx2here, OriginIx2here, Ix2_min, Ix2_max, type_required); + + //return(Descendents (ScanState.label, ScanIx2here, OriginIx2here, Ix2_min, Ix2_max, type_required)); + return(desc_found); + } + + // Specialization for Heisenberg: + + Vect Descendents (const Heis_Bethe_State& ScanState, const Heis_Bethe_State& OriginState, int type_required) + { + return(Descendents (ScanState.label, ScanState.Ix2, OriginState.Ix2, ScanState.base.Ix2_min, ScanState.base.Ix2_max, type_required)); + } + + + + + + + // ABACUS++T version: + /* + //Vect Descendents (const LiebLin_Bethe_State& ScanState, const LiebLin_Bethe_State& OriginState) + Vect Descendents (const LiebLin_Bethe_State& ScanState, const LiebLin_Bethe_State& OriginState, int type_required) + //Descendent_Data Descendents (const LiebLin_Bethe_State& ScanState, const LiebLin_Bethe_State& OriginState, int type_required) + { + + // THESE EXPLANATIONS ARE DEPRECATED + + // A given state is labeled by pairs (Ix2old, Ix2exc) of modified quantum numbers, + // say (Ix2old[0][0], Ix2exc[0][0]) ... (Ix2old[0][nexc[0] ], Ix2exc[0][nexc[0] ]) at level 0 + // with the constraint that Ix2old and Ix2exc are strictly ordered. + // Higher levels are treated similarly. + + // By definition, the descendents are constructed according to two possibilities: + + // First, if type_required == 0, by taking the inner(right)most index particle (so at index (nexc[i]+1)/2) + // and moving it one step further in the direction it was moving, if possible. + // The expected data_value in this case is simply data_value_ref of the ancestor. + + // Second, if type_required == 1, by adding another Ix2old Ix2exc pair, + // letting the hole take all possible values of Ix2old between Ix2old[i][(nexc[i]+1)/2 ] and Ix2old[i][(nexc[i]+1)/2 + 1] + // and putting the corresponding Ix2exc to the first available position on the left or right (if any), + // without annihilating another hole. In other words, we add a Ix2old Ix2exc pair in the (right)middle, respecting the ordering rule. + // The expected data_value in this case is the ancestor's data_value multiplied by the cost of adding a particle-hole (ph_cost). + + // IMPORTANT NOTE: THE IMPLEMENTATION IS FOR NOW ONLY VALID FOR LIEB-LINIGER. + // The logic is however extensible to the case of multiple base levels. + + // Number of descendents: + int ndesc_possible = (type_required == 0 ? 1 : ScanState.N * 2); // upper bound guess; should be refined. + Vect desclabelfound (ndesc_possible); + Vect desctypefound (ndesc_possible); + int ndesc_found = 0; + + // First read the current state data: + State_Label_Data currentdata = Read_State_Label (ScanState.label, OriginState.Ix2); + + if (type_required == 0) { + // We move the inner(right)most hole one more step towards the center of the block of Ix2 in which it sits. + + // Is there already an excitation at level 0? If so, move inner(right)most particle one step further. + if (currentdata.nexc[0] > 0) { + + //cout << "Here a" << endl; + + // Produce a new state_label_data: + State_Label_Data descdata = Read_State_Label (ScanState.label, OriginState.Ix2); + // Identify the inner(right)most excitation: + int innerindex = currentdata.nexc[0]/2; + + int newholepos = currentdata.Ix2old[0][innerindex]; + + // Determine the size of the block of OriginState.Ix2 in which this hole sits: + int nroccupiedtoleft = 0; + while (OriginState.Ix2.includes (currentdata.Ix2old[0][innerindex] -2*(nroccupiedtoleft + 1))) + //while (ScanState.Ix2.includes (currentdata.Ix2old[0][innerindex] -2*(nroccupiedtoleft + 1))) + //while (OriginState.Ix2.includes (currentdata.Ix2old[0][innerindex] -2*(nroccupiedtoleft + 1)) + //&& ScanState.Ix2.includes (currentdata.Ix2old[0][innerindex] -2*(nroccupiedtoleft + 1))) + ////&& !currentdata.Ix2old[0].includes(currentdata.Ix2old[0][innerindex] -2*(nroccupiedtoleft + 1))) + nroccupiedtoleft++; + int nroccupiedtoright = 0; + while (OriginState.Ix2.includes (currentdata.Ix2old[0][innerindex] +2*(nroccupiedtoright + 1))) + //while (ScanState.Ix2.includes (currentdata.Ix2old[0][innerindex] +2*(nroccupiedtoright + 1))) + //while (OriginState.Ix2.includes (currentdata.Ix2old[0][innerindex] +2*(nroccupiedtoright + 1)) + // && ScanState.Ix2.includes (currentdata.Ix2old[0][innerindex] +2*(nroccupiedtoright + 1))) + ////&& !currentdata.Ix2old[0].includes(currentdata.Ix2old[0][innerindex] +2*(nroccupiedtoright + 1))) + nroccupiedtoright++; + + //cout << "Here b" << endl; + + //cout << "holeIx2 = " << currentdata.Ix2old[0][innerindex] << "\tnroccupiedtoleft = " << nroccupiedtoleft << "\tnroccupiedtoright = " << nroccupiedtoright << endl; + // Move the hole further in: + // Requirements for end configuration: hole must be further near middle, so + // if nroccupiedtoleft < nroccupiedtoright, move towards the right. + // if nroccupiedtoleft > nroccupiedtoright, move towards the left. + // If the two are then equal, choose hole moved from right to left. + // if moved towards the right: nroccupiedtoleft + 1 < nroccupiedtoright - 1 + // if moved towards the left: nroccupiedtoleft - 1 <= nroccupiedtoright + 1 + + if (//nroccupiedleft < nroccupiedtoright && // redundant from next condition + nroccupiedtoleft - nroccupiedtoright + 2 < 0) newholepos += 2; + else if (// nroccupiedright > nroccupiedtoright && // redundant from next condition + nroccupiedtoleft - nroccupiedtoright - 2 >= 0) newholepos -= 2; + + if (newholepos != currentdata.Ix2old[0][innerindex] // we have successfully moved the hole + && !currentdata.Ix2old[0].includes(newholepos) // new hole position is not already taken + ) + { // we have found a descendent + descdata.Ix2old[0][innerindex] = newholepos; + desclabelfound[ndesc_found] = Return_State_Label (descdata, OriginState.Ix2); + desctypefound[ndesc_found++] = 0; + } + + //cout << "Here c" << endl; + + } // if (nexc > 0) + } // if (type_required == 0) + + + if (type_required == 1) { + // Is there already an excitation at level 0? If so, move inner(right)most particle one step further. + if (currentdata.nexc[0] > 0) { + // Produce a new state_label_data: + State_Label_Data descdata = Read_State_Label (ScanState.label, OriginState.Ix2); + // Identify the inner(right)most excitation: + int innerindex = currentdata.nexc[0]/2; + + int newpartpos = currentdata.Ix2exc[0][innerindex]; + + // Determine the size of the vacancy block of OriginState.Ix2 in which this particle sits: + int nremptytoleft = 0; + while (!OriginState.Ix2.includes (currentdata.Ix2exc[0][innerindex] -2*(nremptytoleft+1))) + nremptytoleft++; + int nremptytoright = 0; + while (!OriginState.Ix2.includes (currentdata.Ix2exc[0][innerindex] + 2*(nremptytoright+1))) + nremptytoright++; + // BUG HERE: doesn't terminate if there is no particle to the left or right... + + + if (nremptytoleft - nremptytoright + 2 < 0) newpartpos += 2; + else if (nremptytoleft - nremptytoright + 2 >= 0) newpartpos -= 2; + + if (newpartpos != currentdata.Ix2exc[0][innerindex] // we have successfully moved the hole + && !currentdata.Ix2exc[0].includes(newpartpost) // new particle position is not already taken + ) + { // we have found a descendent + descdata.Ix2exc[0][innerindex] = newpartpos; + desclabelfound[ndesc_found] = Return_State_Label (descdata, OriginState.Ix2); + desctypefound[ndesc_found++] = 1; + } + + // FROM ++T, DEPRECATED IN ++G: + // Can this Ix2exc be moved one more step in the same direction? Try it out. + ////descdata.Ix2exc[0][innerindex] += (currentdata.Ix2exc[0][innerindex] > currentdata.Ix2old[0][innerindex] ? 2 : -2); + //int newpartpos = currentdata.Ix2exc[0][innerindex] + (currentdata.Ix2exc[0][innerindex] > currentdata.Ix2old[0][innerindex] ? 2 : -2); + // Move until we get a consistent new configuration: + // First, this quantum number was not already occupied + // Second, this quantum number does not annihilate one of the holes: + //while (ScanState.Ix2.includes(newpartpos) || currentdata.Ix2old[0].includes(newpartpos)) + // newpartpos += (currentdata.Ix2exc[0][innerindex] > currentdata.Ix2old[0][innerindex] ? 2 : -2); + + // If this new particle position is to the right of the Ix2exc of one index lower, we've got a descendent: + //if (innerindex == 0 || newpartpos > currentdata.Ix2exc[0][innerindex - 1] + // //&& abs(newpartpos) <= 13 // TEMPORARY LIMIT: INDUCE FINITE NR OF STATES + // && (innerindex == currentdata.nexc[0] - 1 || newpartpos < currentdata.Ix2exc[0][innerindex + 1])) { + // We have a descendent, + ////cout << "\tFound increased displacement at innerindex " << innerindex << endl; + ////cout << descdata.Ix2exc[0][innerindex] << " to " << newpartpos << endl; + //descdata.Ix2exc[0][innerindex] = newpartpos; + //desclabelfound[ndesc_found] = Return_State_Label (descdata, OriginState.Ix2); + //desctypefound[ndesc_found++] = 1; + //} + + } //if (currentdata.nexc[0] > 0) + } // if (type_required == 1) + + + if (type_required == 2) { + // Now add a new p-h pair at the inner(right)most position, scanning the hole position immediately + // but putting the particle to the first available left and right positions (if any). + + // type_required == 2 means that the hole must be either at the edge of an OriginIx2 domain, or next to an existing hole. + + // Define the objects for the newstatedata: + Vect type_new = currentdata.type; + Vect M_new = currentdata.M; + Vect nexc_new = currentdata.nexc; + nexc_new[0] += 1; // we drill one more particle-hole pair at level 0 + int index_new = (currentdata.nexc[0] + 1)/2; // we put the new p-h pair at index index_new. + + int ntypespresent = 1; // only one type for LiebLin + Vect > Ix2old_new(ntypespresent); + Vect > Ix2exc_new(ntypespresent); + for (int it = 0; it < ntypespresent; ++it) Ix2old_new[it] = Vect(JSC::max(nexc_new[it],1)); + for (int it = 0; it < ntypespresent; ++it) Ix2exc_new[it] = Vect(JSC::max(nexc_new[it],1)); + + // Copy earlier data in: + for (int it = 0; it < ntypespresent; ++it) { + for (int i = 0; i < currentdata.nexc[it]; ++i) { + Ix2old_new[it][i + (i >= index_new)] = currentdata.Ix2old[it][i]; + Ix2exc_new[it][i + (i >= index_new)] = currentdata.Ix2exc[it][i]; + } + //cout << "current Ix2old: " << currentdata.Ix2old[0] << endl; + //cout << "current Ix2exc: " << currentdata.Ix2exc[0] << endl; + //cout << "allocated new Ix2old: " << Ix2old_new[0] << endl; + //cout << "allocated new Ix2exc: " << Ix2exc_new[0] << endl; + } + + State_Label_Data descdatanewph (type_new, M_new, nexc_new, Ix2old_new, Ix2exc_new); + + //cout << "\ttype: " << descdatanewph.type << endl; + //cout << "\tM: " << descdatanewph.M << endl; + //cout << "\tnexc: " << descdatanewph.nexc << endl; + //cout << "\tIx2old: " << descdatanewph.Ix2old[0] << endl; + //cout << "\tIx2new: " << descdatanewph.Ix2exc[0] << endl; + + // Now find all the possible positions for the new hole: + // The new hole should be between Ix2old_new[0][innerindex - 1] and Ix2old_new[0][innerindex + 1] + // if those exist (i.e., modulo lack of constraint at the left/right boundaries). + int newpartpos; + for (int i = 0; i < ScanState.N; ++i) { + + // Determine the size of the block of OriginState.Ix2 in which this hole sits: + int nroccupiedtoleft = 0; + while (OriginState.Ix2.includes (ScanState.Ix2[i] -2*(nroccupiedtoleft + 1))) + nroccupiedtoleft++; + int nroccupiedtoright = 0; + while (OriginState.Ix2.includes (ScanState.Ix2[i] +2*(nroccupiedtoright + 1))) + nroccupiedtoright++; + // This will be used in the next conditions to check whether the new hole would sit in a left or right block. + + if ((currentdata.nexc[0] == 0 // no earlier p-h in state so no left boundary for new hole; index_new == 0 + || ScanState.Ix2[i] > currentdata.Ix2old[0][index_new - 1]) + && (currentdata.nexc[0] <= 1 // at most one p-h in state so no right boundary for new hole; index_new <= 1 + || ScanState.Ix2[i] < currentdata.Ix2old[0][index_new]) + //&& (currentdata.nexc[0] == 0 || !currentdata.Ix2exc[0].includes(ScanState.Ix2[i])) + // new hole is not one of the already existing p excitations // FULFILLED BY ABOVE CONDITIONS IF EXC ARE ORDERED + //&& (!ScanState.Ix2.includes(ScanState.Ix2[i] - 2) || !ScanState.Ix2.includes(ScanState.Ix2[i] + 2)) + && ( + // New hole must be created on the boundary of an OriginState Ix2 domain, + // or next to a preexisting hole in the domain. + (!OriginState.Ix2.includes(ScanState.Ix2[i] - 2)) + // new hole is on the left boundary of an OriginState Ix2 domain. + || !OriginState.Ix2.includes(ScanState.Ix2[i] + 2) + // new hole is on the right boundary of an OriginState Ix2 domain. + //|| (currentdata.nexc[0] > 0 && ScanState.Ix2[i] == currentdata.Ix2old[0][index_new - 1] + 2) + || type_required == 2 && (currentdata.nexc[0] > 0 && nroccupiedtoleft < nroccupiedtoright + && ScanState.Ix2[i] == currentdata.Ix2old[0][index_new - 1] + 2) + // new hole is in a left domain and immediately to the right of first hole to its left + //|| (currentdata.nexc[0] > 1 && ScanState.Ix2[i] == currentdata.Ix2old[0][index_new] - 2) + || type_required == 2 && (currentdata.nexc[0] > 1 && nroccupiedtoleft >= nroccupiedtoright + && ScanState.Ix2[i] == currentdata.Ix2old[0][index_new] - 2) + // new hole is in a right domain and immediately to the left of first hole to its right + ) + ) { + + bool acceptable_hole = true; + + if (currentdata.nexc[0] > 0 && currentdata.Ix2exc[0].includes(ScanState.Ix2[i])) acceptable_hole = false; + + // Determine the size of the block of Ix2 in which this hole sits: + int nroccupiedtoleft = 0; + while (OriginState.Ix2.includes (ScanState.Ix2[i] -2*(nroccupiedtoleft + 1))) + nroccupiedtoleft++; + int nroccupiedtoright = 0; + while (OriginState.Ix2.includes (ScanState.Ix2[i] +2*(nroccupiedtoright + 1))) + nroccupiedtoright++; + + // The hole is unacceptable if it breaks the `towards the block center' rule: + if (nroccupiedtoleft < nroccupiedtoright // new hole would be in left part of block + && OriginState.Ix2.includes(ScanState.Ix2[i] - 2) // and is not at the boundary + && currentdata.nexc[0] > 0 && ScanState.Ix2[i] != currentdata.Ix2old[0][index_new - 1] + 2) + // and it's not immediately to the left of a preexisting hole + acceptable_hole = false; + if (nroccupiedtoleft >= nroccupiedtoright // new hole could be in right part of block + && OriginState.Ix2.includes(ScanState.Ix2[i] + 2) // and is not at the boundary + && currentdata.nexc[0] > 2 && ScanState.Ix2[i] != currentdata.Ix2old[0][index_new] - 2) + // and it's not immediately to the right of a preexisting hole + acceptable_hole = false; + + if (acceptable_hole) { + //cout << "Found a possible hole position at index " << i << " and Ix2 " << ScanState.Ix2[i] << endl; + + // ScanState.Ix2[i] is an allowable new hole position. + + + // We now look for all possible particle positions, + // the allowable ones being either at the edge of a block, or next to an existing particle: + + + // Find the first available particle position to the left: + newpartpos = ScanState.Ix2[i] - 2; + // This must be lower than the Ix2exc of one index higher (if any): + if (currentdata.nexc[0] >= 2) newpartpos = JSC::min(newpartpos, currentdata.Ix2exc[0][index_new] - 2); + // The new particle position must not already be occupied in ScanState, + // and must not be at one of the already specified holes + while (ScanState.Ix2.includes(newpartpos) + || (currentdata.nexc[0] > 0 && currentdata.Ix2old[0].includes(newpartpos))) newpartpos -= 2; + // If this new particle position is to the right of the Ix2exc of one index lower (if any), we've got a descendent: + if (currentdata.nexc[0] == 0 || newpartpos > currentdata.Ix2exc[0][index_new - 1]) { + //cout << "\tFound a particle position (moving L) at " << newpartpos << " with index " << index_new << endl; + //if (abs(newpartpos) <= 13) { // TEMPORARY LIMIT: INDUCE FINITE NR OF STATES + descdatanewph.Ix2old[0][index_new] = ScanState.Ix2[i]; + descdatanewph.Ix2exc[0][index_new] = newpartpos; + //cout << "ScanState.Ix2[i] = " << ScanState.Ix2[i] << "\tnewpartpos = " << newpartpos << endl; + //cout << "\tIx2old: " << descdatanewph.Ix2old[0] << endl; + //cout << "\tIx2new: " << descdatanewph.Ix2exc[0] << endl; + desclabelfound[ndesc_found] = Return_State_Label (descdatanewph, OriginState.Ix2); + desctypefound[ndesc_found++] = type_required; + //cout << "\tLabel found = " << desclabelfound[ndesc_found - 1] << endl; + //} // TEMP + } + + + // Now find the first particle position to the right: + newpartpos = ScanState.Ix2[i] + 2; + // This must be higher than the Ix2exc of one index lower (if any): + if (index_new > 0) newpartpos = JSC::max(newpartpos, currentdata.Ix2exc[0][index_new - 1] + 2); + // The new particle position must not already be occupied in ScanState, + // and must not be at one of the already specified holes + while (ScanState.Ix2.includes(newpartpos) + || (currentdata.nexc[0] > 0 && currentdata.Ix2old[0].includes(newpartpos))) newpartpos += 2; + // If this new particle position is to the left of the Ix2exc of one index higher (if any), we've got a descendent: + if (currentdata.nexc[0] <= 1 || newpartpos < currentdata.Ix2exc[0][index_new]) { + //// If this new particle position is to the left of the Ix2exc of one index higher (if any), got descendent: + //if (currentdata.nexc[0] <= 1 || newpartpos < currentdata.Ix2exc[0][index_new]) { + //cout << "\tFound a particle position (moving R) at " << newpartpos << " with index " << index_new << endl; + + //if (abs(newpartpos) <= 13) { // TEMPORARY LIMIT: INDUCE FINITE NR OF STATES + descdatanewph.Ix2old[0][index_new] = ScanState.Ix2[i]; + descdatanewph.Ix2exc[0][index_new] = newpartpos; + //cout << "ScanState.Ix2[i] = " << ScanState.Ix2[i] << "\tnewpartpos = " << newpartpos << endl; + //cout << "\tIx2old: " << descdatanewph.Ix2old[0] << endl; + //cout << "\tIx2new: " << descdatanewph.Ix2exc[0] << endl; + desclabelfound[ndesc_found] = Return_State_Label (descdatanewph, OriginState.Ix2); + desctypefound[ndesc_found++] = type_required; + //cout << "\tLabel found = " << desclabelfound[ndesc_found - 1] << endl; + //} // TEMP + } + } // if (acceptable_hole + } + } // for i + } // if (type_required == 2 || type_required == 3) + + + Vect desclabelfoundresized (ndesc_found); + //Vect_DP desctypefoundresized (ndesc_found); + for (int i = 0; i < ndesc_found; ++i) { + desclabelfoundresized[i] = desclabelfound[i]; + //desctypefoundresized[i] = descvalue[i]; + } + + //for (int i = 0; i < ndesc_found; ++i) if (desclabelfound[i] == "16_2_13@17:15@19") cout << "16_2_13@17:15@19 descended from " << ScanState.label << endl; + + //if (ScanState.label == "16_0_") cout << "State 16_0_: descendents " << desclabelfound << endl; + //if (ScanState.label == "16_1_13@17") cout << "State 16_1_13@17: descendents " << desclabelfound << endl; + + //cout << "Found " << ndesc_found << " descendents, " << desclabelfoundresized << endl; + + //Descendent_Data descdata(); + + //descdata.label = desclabelfoundresized; + //descdata.type = desctypefoundresized; + + return(desclabelfoundresized); + //return(descdata); + } + + + //Vect Descendents (const Heis_Bethe_State& ScanState, const Heis_Bethe_State& OriginState) + Vect Descendents (const Heis_Bethe_State& ScanState, const Heis_Bethe_State& OriginState, int type_required) + //Descendent_Data Descendents (const Heis_Bethe_State& ScanState, const Heis_Bethe_State& OriginState, int type_required); + { + // NOT IMPLEMENTED YET + JSCerror("Descendents for Heis not implemented yet."); + return(Vect()); + } + + */ + +} // namespace JSC + diff --git a/src/SCAN/General_Scan.cc b/src/SCAN/General_Scan.cc new file mode 100644 index 0000000..f272805 --- /dev/null +++ b/src/SCAN/General_Scan.cc @@ -0,0 +1,1813 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/SCAN/General_Scan.cc + +Purpose: universal implementation of state scanning: + functions to descend down hierarchy of intermediate states. + +NOTE: since templated functions have to be in the same file, + we put all scanning functions here. The externally-accessible + functions are defined at the end of this file. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + + +string LABEL_TO_CHECK = "bla"; +//string LABEL_TO_CHECK = "6_0_"; +//string LABEL_TO_CHECK = "6_2_22y32"; + + +namespace JSC { + + // Types of descendents: + // 14 == iK stepped up, leading exc one step further (can lead to ph recombination) + // 13 == iK stepped up, next exc (nr of ph preserved, not taking possible ph recombination into account) + // 12 == iK stepped up, next ext (nr ph increased, not taking possible ph recombination into account) + // 11 == iK stepped down, leading exc one step further (can lead to ph recombination) + // 10 == iK stepped down, next exc (nr of ph preserved, not taking possible ph recombination into account) + // 9 == iK stepped down, next exc (nr ph increased, not taking possible ph recombination into account) + // 8 == iK preserved, 14 and 11 (up to 2 ph recombinations) + // 7 == iK preserved, 14 and 10 + // 6 == iK preserved, 14 and 9 + // 5 == iK preserved, 13 and 11 + // 4 == iK preserved, 13 and 10 + // 3 == iK preserved, 13 and 9 + // 2 == iK preserved, 12 and 11 + // 1 == iK preserved, 12 and 10 + // 0 == iK preserved, 12 and 9 + + // For scanning over symmetric states, the interpretation is slightly different. + // Types 14, 13 and 12 step iK up using the Ix2 on the right only, and mirrors the change on the left Ix2. + // Types 11, 10 and 9 step iK down using the Ix2 on the right only, and mirrors the change on the left Ix2. + // There is then no need of scanning over types 0 - 8. + // By convention, types 9, 10 and 11 can call types 9 - 14; types 12-14 can only call types 12-14. + + bool Expect_ph_Recombination_iK_Up (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2) + { + // This function returns true if descending further can lead to a particle-hole recombination. + // The criteria which are used are: + // - the active excitation has moved at least one step (so it has already created its p-h pair) + // - there exists an OriginIx2 between the active Ix2 and the next Ix2 (to right or left depending on type of descendent) + + Vect > ScanIx2 = Return_Ix2_from_Label (ScanIx2_label, OriginIx2); + + // Determine the level and index of the bottom-most left-most right-moving quantum number sits: + int exclevel = -1; + int excindex = 0; + bool excfound = false; + do { + exclevel++; + if (exclevel == ScanIx2.size()) { // there isn't a single right-moving quantum number in ScanIx2 + break; + } + for (int alpha = 0; alpha < ScanIx2[exclevel].size(); ++alpha) + if (ScanIx2[exclevel][alpha] > BaseScanIx2[exclevel][alpha]) { + excindex = alpha; + excfound = true; + break; + } + } while (!excfound); + // If we haven't found an excitation, then exclevel == ScanIx2.size() and excindex = 0; + + if (excfound && !BaseScanIx2[exclevel].includes(ScanIx2[exclevel][excindex])) { // there exists an already dispersing excitation which isn't in Origin + // Is there a possible recombination? + if (excindex < ScanIx2[exclevel].size() - 1) { // a particle to the right of excitation has already move right, so there is a hole + // check that there exists an occupied Ix2 in Origin sitting between the excitation and the next Ix2 to its right in ScanIx2 + for (int alpha = BaseScanIx2[exclevel].size() - 1; alpha >= 0; --alpha) + if (BaseScanIx2[exclevel][alpha] > ScanIx2[exclevel][excindex] && BaseScanIx2[exclevel][alpha] < ScanIx2[exclevel][excindex + 1]) { + return(true); + } + } + } // if (excfound) + + return(false); + } + + // Specialization for Lieb-Liniger: + bool Expect_ph_Recombination_iK_Up (string ScanIx2_label, const LiebLin_Bethe_State& OriginState) + { + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginState.Ix2; + Vect > BaseScanIx2here(1); + BaseScanIx2here[0] = OriginState.Ix2; + return(Expect_ph_Recombination_iK_Up (ScanIx2_label, OriginIx2here, BaseScanIx2here)); + } + + // Specialization for Heis + bool Expect_ph_Recombination_iK_Up (string ScanIx2_label, const Heis_Bethe_State& OriginState) + { + return(Expect_ph_Recombination_iK_Up (ScanIx2_label, OriginState.Ix2, OriginState.Ix2)); + } + + + + bool Expect_ph_Recombination_iK_Down (string ScanIx2_label, const Vect >& OriginIx2, const Vect >& BaseScanIx2) + { + // This function returns true if descending further can lead to a particle-hole recombination. + // The criteria which are used are: + // - the active excitation has moved at least one step (so it has already created its p-h pair) + // - there exists an OriginIx2 between the active Ix2 and the next Ix2 (to right or left depending on type of descendent) + + Vect > ScanIx2 = Return_Ix2_from_Label (ScanIx2_label, OriginIx2); + + // Determine the level and index of the bottom-most right-most left-moving quantum number sits: + int exclevel = -1; + int excindex = 0; + bool excfound = false; + + //cout << "Looking for exclevel and excindex for " << endl << "\tBaseIx2 = " << BaseScanIx2 << endl << "\tScanIx2 = " << ScanIx2 << endl; + do { + exclevel++; + if (exclevel == ScanIx2.size()) { // there isn't a single left-moving quantum number in ScanIx2 + break; + } + for (int alpha = ScanIx2[exclevel].size() - 1; alpha >= 0; --alpha) { + //cout << exclevel << "\t" << alpha << "\t" << ScanIx2[exclevel][alpha] << "\t" << BaseScanIx2[exclevel][alpha] << "\t" << (ScanIx2[exclevel][alpha] < BaseScanIx2[exclevel][alpha]) << endl; + if (ScanIx2[exclevel][alpha] < BaseScanIx2[exclevel][alpha]) { + excindex = alpha; + excfound = true; + break; + } + } + } while (!excfound); + // If we haven't found an excitation, then exclevel == ScanIx2.size() and excindex = 0; + if (!excfound) excindex = ScanIx2[exclevel].size() - 1; + + if (excfound && !BaseScanIx2[exclevel].includes(ScanIx2[exclevel][excindex])) { // there exists an already dispersing excitation which isn't in Origin + // Is there a possible recombination? + if (excindex > 0) { // a particle to the left of excitation has already moved left, so there is a hole + // check that there exists an occupied Ix2 in Origin sitting between the excitation and the next Ix2 to its left in ScanIx2 + for (int alpha = 0; alpha < BaseScanIx2[exclevel].size(); ++alpha) + if (BaseScanIx2[exclevel][alpha] > ScanIx2[exclevel][excindex - 1] && BaseScanIx2[exclevel][alpha] < ScanIx2[exclevel][excindex]) { + return(true); + } + } + } // if (excfound) + + return(false); + } + + // Specialization for Lieb-Liniger: + bool Expect_ph_Recombination_iK_Down (string ScanIx2_label, const LiebLin_Bethe_State& OriginState) + { + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginState.Ix2; + Vect > BaseScanIx2here(1); + BaseScanIx2here[0] = OriginState.Ix2; + return(Expect_ph_Recombination_iK_Down (ScanIx2_label, OriginIx2here, BaseScanIx2here)); + } + + // Specialization for Heis + bool Expect_ph_Recombination_iK_Down (string ScanIx2_label, const Heis_Bethe_State& OriginState) + { + return(Expect_ph_Recombination_iK_Down (ScanIx2_label, OriginState.Ix2, OriginState.Ix2)); + } + + + + /* + template + void Descend_and_Compute_for_Fixed_Base (char whichDSF, Tstate& AveragingState, Tstate& BaseScanState, Tstate& ScanState, + int type_required, int iKmin, int iKmax, int iKmod, + //Scan_Thread_List& paused_thread_list, + //Scan_Thread_Set& paused_thread_set, + Scan_Thread_Data& paused_thread_data, + //thresholdremoved DP& running_scan_threshold, //DP ref_abs_data_value, + DP& ph_cost, int Max_Secs, DP sumrule_factor, DP Chem_Pot, Scan_Info& scan_info, + fstream& RAW_outfile, fstream& INADM_outfile, int& ninadm, + fstream& CONV0_outfile, int& nconv0, fstream& STAT_outfile) + { + + //cout << "Calling descent with type_required " << type_required << " on state " << ScanState.label << "\t" << Return_Ix2_from_Label(ScanState.label, AveragingState.Ix2) << endl; + //cout << "Calling descent with type_required " << type_required << " on state " << ScanState.label << endl; + ScanState.Compute_Momentum(); + Vect desc_label; + + // ++G_7 logic + bool disperse_only_current_exc_up = false; + if (type_required == 14 || type_required == 8 || type_required == 7 || type_required == 6) disperse_only_current_exc_up = true; + bool preserve_nexc_up = false; + if (type_required == 13 || type_required == 5 || type_required == 4 || type_required == 3) preserve_nexc_up = true; + bool disperse_only_current_exc_down = false; + if (type_required == 11 || type_required == 8 || type_required == 5 || type_required == 2) disperse_only_current_exc_down = true; + bool preserve_nexc_down = false; + if (type_required == 10 || type_required == 7 || type_required == 4 || type_required == 1) preserve_nexc_down = true; + + if (whichDSF == 'B') { // symmetric state scanning + if (type_required >= 9 && type_required <= 11) + desc_label = Descendent_States_with_iK_Stepped_Down_rightIx2only (ScanState.label, BaseScanState, disperse_only_current_exc_down, preserve_nexc_down); + else if (type_required >= 12 && type_required <= 14) + desc_label = Descendent_States_with_iK_Stepped_Up_rightIx2only (ScanState.label, BaseScanState, disperse_only_current_exc_up, preserve_nexc_up); + } + else { + if (type_required >= 0 && type_required <= 8) { + desc_label = Descendent_States_with_iK_Preserved(ScanState.label, BaseScanState, disperse_only_current_exc_up, preserve_nexc_up, disperse_only_current_exc_down, preserve_nexc_down); + } + else if (type_required >= 9 && type_required <= 11) + desc_label = Descendent_States_with_iK_Stepped_Down (ScanState.label, BaseScanState, disperse_only_current_exc_down, preserve_nexc_down); + else if (type_required >= 12 && type_required <= 14) + desc_label = Descendent_States_with_iK_Stepped_Up (ScanState.label, BaseScanState, disperse_only_current_exc_up, preserve_nexc_up); + } + //cout << "Found " << desc_label.size() << " descendents." << endl; + //for (int is = 0; is < desc_label.size(); ++is) cout << "is " << is << "\tdesc: " << desc_label[is] << "\t" << Return_Ix2_from_Label(desc_label[is], AveragingState.Ix2) << endl; + //char a; + //cin >> a; + + //cout << "OK for descend on " << ScanState.label << " with type_required = " << type_required << endl; + //cout << desc_label << endl; + + //Vect desc_label = desc_data.label; + //Vect desc_type = desc_data.type; + + //bool disp_OK = false; + + //if (ScanState.label == "7|2:1_0|1_|0@2") { + //if (ScanState.label == "64_1_63@319") { + //if (ScanState.label == "32_1_-29@-33") { + + if (ScanState.label == LABEL_TO_CHECK) { + cout << "Called Descend on state " << ScanState << endl; + cout << "For type_required == " << type_required << ", found " << desc_label.size() << " descendents, "; + for (int i = 0; i < desc_label.size(); ++i) { + //cout << desc_label[i] << "\t"; cout << endl; + ScanState.Set_to_Label (desc_label[i], BaseScanState.Ix2); + ScanState.Compute_All(true); + cout << ScanState << endl; + } + cout << "Do you want to follow one of these descendents? (y/n)" << endl; + char a; cin >> a; + if (a == 'y') { + cout << "Which label do you want to follow?" << endl; + cin >> LABEL_TO_CHECK; + } + } + + + string label_here = ScanState.label; + //int ScanState_iK = ScanState.iK; + + for (int idesc = 0; idesc < desc_label.size(); ++idesc) { + + //cout << "\tDealing with descendent " << idesc << " out of " << desc_label.size() << " with label " << desc_label[idesc] << endl; + //cout << "\tfrom state with label " << label_here << " and of type_required " << type_required << endl; + + clock_t start_time_here = clock(); + + //if (desc_label[idesc] == "64_2_0yvv7") { + if (false) { + cout << "Found " << desc_label[idesc] << " as descendent of type " << type_required << " of " << label_here << endl; + ScanState.Set_to_Label (label_here, BaseScanState.Ix2); + cout << ScanState.Ix2 << endl; + //cout << "Found " << desc_label.size() << " descendents, " << endl; + //for (int i = 0; i < desc_label.size(); ++i) cout << desc_label[i] << "\t"; cout << endl; + ScanState.Set_to_Label (desc_label[idesc], BaseScanState.Ix2); + cout << ScanState.Ix2 << endl; + //ScanState.Compute_All(true); + //cout << "Resulting Ix2: " << ScanState.Ix2 << endl; + //cout << ScanState << endl; + //cout << "Admissible: " << ScanState.Check_Admissibility(whichDSF) << endl; + //char a; cin >> a; + } + + + ScanState.Set_to_Label (desc_label[idesc], BaseScanState.Ix2); + + bool admissible = ScanState.Check_Admissibility(whichDSF); + + DP data_value = 0.0; + + scan_info.Ndata++; + + ScanState.conv = false; + ScanState.Compute_Momentum(); // since momentum is used as forced descent criterion + + + if (admissible) { + + ScanState.Compute_All (idesc == 0); + //ScanState.Compute_All (true); + + //scan_info.Ndata++; + + if (ScanState.conv) { + scan_info.Ndata_conv++; + + // Put momentum in fundamental window, if possible: + int iKexc = ScanState.iK - AveragingState.iK; + while (iKexc > iKmax && iKexc - iKmod >= iKmin) iKexc -= iKmod; + while (iKexc < iKmin && iKexc + iKmod <= iKmax) iKexc += iKmod; + + data_value = Compute_Matrix_Element_Contrib (whichDSF, iKmin, iKmax, ScanState, AveragingState, Chem_Pot, RAW_outfile); + if (iKexc >= iKmin && iKexc <= iKmax) scan_info.sumrule_obtained += data_value*sumrule_factor; + //cout << "data_value found = " << data_value * sumrule_factor << endl; + + // Uncomment line below if .stat file is desired: + //STAT_outfile << setw(20) << label_here << "\t" << setw(5) << type_required << "\t" << setw(16) << std::scientific << running_scan_threshold << "\t" << setw(20) << ScanState.label << "\t" << setw(16) << data_value << "\t" << setw(16) << std::fixed << setprecision(8) << data_value/running_scan_threshold << endl; + + } // if (ScanState.conv) + else { + if (nconv0++ < 1000) + CONV0_outfile << setw(25) << ScanState.label << setw(25) << ScanState.diffsq << setw(5) << ScanState.Check_Rapidities() + << setw(25) << ScanState.String_delta() << endl; + scan_info.Ndata_conv0++; + //cout << "State did not converge." << endl; + } + } // if (admissible) + + else { + if (ninadm++ < 1000000) INADM_outfile << ScanState.label << endl; + scan_info.Ninadm++; + //cout << "State was inadmissible." << endl; + // Set data_value to enable continued scanning later on: + //thresholdremoved data_value = 0.1* running_scan_threshold; + } + + clock_t stop_time_here = clock(); + + scan_info.CPU_ticks += stop_time_here - start_time_here; + + Tstate state_to_descend; state_to_descend = ScanState; // for checking + + ScanState.Compute_Momentum(); + // Put momentum in fundamental window, if possible: + int iKexc = ScanState.iK - AveragingState.iK; + while (iKexc > iKmax && iKexc - iKmod >= iKmin) iKexc -= iKmod; + while (iKexc < iKmin && iKexc + iKmod <= iKmax) iKexc += iKmod; + + // ++G_7 logic + // Momentum-preserving are only descended to momentum-preserving. + // Momentum-increasing are only descended to momentum-preserving and momentum-increasing. + // Momentum-decreasing are only descended to momentum-preserving and momentum-decreasing. + Vect allowed(false, 15); + if (whichDSF == 'B') { + // We scan over symmetric states. Only types 14 down to 9 are allowed. + if (type_required >= 9 && type_required <= 11) { // iK stepped down on rightIx2; step further up or down + allowed[9] = true; allowed[10] = true; allowed[11] = true; + allowed[12] = true; allowed[13] = true; allowed[14] = true; + } + else if (type_required >= 12 && type_required <= 14) { // iK stepped up on rightIx2; only step further up + allowed[12] = true; allowed[13] = true; allowed[14] = true; + } + } + else { + if (type_required >= 0 && type_required <= 8) { // momentum-preserving + allowed[0] = (iKexc >= iKmin && iKexc <= iKmax); + allowed[9] = false; + allowed[12] = false; + } + if (type_required >= 9 && type_required <= 11) { // momentum-decreasing + allowed[0] = (iKexc >= iKmin && iKexc <= iKmax); + allowed[9] = (iKexc > iKmin); + allowed[12] = false; + } + if (type_required >= 12 && type_required <= 14) { // momentum-increasing + allowed[0] = (iKexc >= iKmin && iKexc <= iKmax); + allowed[9] = false; + allowed[12] = (iKexc < iKmax); + } + // The others are just copies of the ones above: + allowed[1] = allowed[0]; allowed[2] = allowed[0]; allowed[3] = allowed[0]; allowed[4] = allowed[0]; allowed[5] = allowed[0]; allowed[6] = allowed[0]; allowed[7] = allowed[0]; allowed[8] = allowed[0]; + allowed[10] = allowed[9]; allowed[11] = allowed[9]; + allowed[13] = allowed[12]; allowed[14] = allowed[12]; + } + + + for (int type_required_here = 0; type_required_here < 15; ++type_required_here) { + + if (!allowed[type_required_here]) continue; + + // Reset ScanState to what it was, if change on first pass + if (type_required_here > 0) ScanState = state_to_descend; + + // We determine if we carry on scanning based on the data_value obtained, or forcing conditions: + // Forcing conditions: + //if (!admissible || Force_Descent(whichDSF, ScanState, AveragingState, type_required_here, iKmod, Chem_Pot)) + ////data_value = 1.01 * running_scan_threshold/ph_cost; // force for all types of desc + //data_value = 1.01 * running_scan_threshold; // only force for no new ph pairs + ////data_value = 1.0; // force for all types of desc + // If we're sitting out of the iKmin & iKmax window, stop: + //if (iKmin != iKmax && (ScanState.iK - AveragingState.iK > iKmax || ScanState.iK - AveragingState.iK < iKmin)) data_value = 0.0; + + //if (abs(data_value) * (type_required_here != 2 ? 1.0 : ph_cost) > running_scan_threshold + //if ((abs(data_value) > running_scan_threshold + //|| Nr_ph_Recombinations_Possible (ScanState.label, BaseScanState, type_required_here) > 0) + + //DP expected_abs_data_value = abs(data_value)/pow(ph_cost, DP(Nr_ph_Recombinations_Possible (ScanState.label, BaseScanState, type_required_here))); + DP expected_abs_data_value = abs(data_value); + + //++G_7 logic + if ((type_required_here == 14 || type_required_here == 8 || type_required_here == 7 || type_required_here == 6) + && Expect_ph_Recombination_iK_Up (ScanState.label, BaseScanState)) expected_abs_data_value /= ph_cost; + if (type_required_here == 12 || type_required_here == 2 || type_required_here == 1 || type_required_here == 0) + expected_abs_data_value *= ph_cost; + if ((type_required_here == 11 || type_required_here == 8 || type_required_here == 5 || type_required_here == 2) + && Expect_ph_Recombination_iK_Down (ScanState.label, BaseScanState)) expected_abs_data_value /= ph_cost; + if (type_required_here == 9 || type_required_here == 6 || type_required_here == 3 || type_required_here == 0) + expected_abs_data_value *= ph_cost; + + //cout << "\tIncluding thread " << expected_abs_data_value << "\t" << ScanState.label << "\t" << type_required_here << endl; + //paused_thread_set.Include_Thread (expected_abs_data_value, ScanState.label, type_required_here); + paused_thread_data.Include_Thread (expected_abs_data_value, ScanState.label, type_required_here); + //cout << "\tDone including thread." << endl; + } // for type_required_here + + //cout << "\tFinished with descendent " << idesc << " out of " << desc_label.size() << " with label " << desc_label[idesc] << endl; + //cout << "\tfrom state with label " << label_here << endl; + } // for idesc + + //cout << "Finished Descend on state " << endl << ScanState.label << endl; + + return; + } + */ + + template + Scan_Info General_Scan (char whichDSF, int iKmin, int iKmax, int iKmod, DP kBT, Tstate& AveragingState, Tstate& SeedScanState, + string defaultScanStatename, int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors) + { + // Performs the scan over excited states, writing data to file. + + // AveragingState is the state on which the correlations are calculated. + // SeedScanState is the originator of all scan states. + // This distinction is kept to allow for quenches and finite temperatures. + + // This function is also called by the parallel implementation of ABACUS. + // In this case, file names carry a rank and nr_processors suffix. + // In fact, the parallelization can be done in incremental levels. + // If paralevel == 0, the run is serial. + // If paralevel == n, the run is parallelized in a tree with n levels of branching. + // A paralevel == 1 branching's files have a suffix of the form "_3_8", meaning that this is the rank 3 out of 8 processors. + // A paralevel == 2 branching's files have a suffix of the form "_3_8_2_8", meaning that this is the rank 2 out of 8 subscan of the _3_8 scan. + + //clock_t start_time = clock(); + //clock_t current_time = clock(); + + //bool in_parallel = (nr_processors > 1); + bool in_parallel = (paralevel > 0); + if (in_parallel && (rank.size() != paralevel || nr_processors.size() != paralevel)) { + cout << "paralevel = " << paralevel << "\trank.size() = " << rank.size() << "\tnr_processors.size() = " << nr_processors.size() << endl; + cout << "rank = " << rank << endl; + cout << "nr_processors = " << nr_processors << endl; + JSCerror("Inconsistent paralevel, rank or nr_processors in General_Scan."); + } + + if (in_parallel && !refine) JSCerror("Must refine when using parallel ABACUS++"); + + DP ph_cost = Particle_Hole_Excitation_Cost (whichDSF, AveragingState); // expected cost on data_value of adding a particle-hole excitation. + + int Max_Secs_used = int(0.9 * Max_Secs); // we don't start any new ithread loop beyond this point + int Max_Secs_alert = int(0.95 * Max_Secs); // we break any ongoing ithread loop beyond this point + + //clock_t start_time_local = clock(); + + stringstream filenameprefix; + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, kBT, AveragingState, SeedScanState, defaultScanStatename); + + if (in_parallel) for (int r = 0; r < paralevel; ++r) filenameprefix << "_" << rank[r] << "_" << nr_processors[r]; + + string prefix = filenameprefix.str(); + + stringstream filenameprefix_prevparalevel; // without the rank and nr_processors of the highest paralevel + Data_File_Name (filenameprefix_prevparalevel, whichDSF, iKmin, iKmax, kBT, AveragingState, SeedScanState, defaultScanStatename); + if (in_parallel) for (int r = 0; r < paralevel - 1; ++r) filenameprefix << "_" << rank[r] << "_" << nr_processors[r]; + + string prefix_prevparalevel = filenameprefix_prevparalevel.str(); + + stringstream RAW_stringstream; string RAW_string; + stringstream INADM_stringstream; string INADM_string; + stringstream CONV0_stringstream; string CONV0_string; + stringstream STAT_stringstream; string STAT_string; + stringstream LOG_stringstream; string LOG_string; + stringstream THR_stringstream; string THR_string; + stringstream THRDIR_stringstream; string THRDIR_string; + stringstream SRC_stringstream; string SRC_string; + //stringstream FSR_stringstream; string FSR_string; + stringstream SUM_stringstream; string SUM_string; + //stringstream SUM_prevparalevel_stringstream; string SUM_prevparalevel_string; + + RAW_stringstream << prefix << ".raw"; + INADM_stringstream << prefix << ".inadm"; + CONV0_stringstream << prefix << ".conv0"; + STAT_stringstream << prefix << ".stat"; + LOG_stringstream << prefix << ".log"; + THR_stringstream << prefix << ".thr"; + THRDIR_stringstream << prefix << "_thrdir"; + SRC_stringstream << prefix << ".src"; + //FSR_stringstream << prefix << ".fsr"; + SUM_stringstream << prefix << ".sum"; + //SUM_prevparalevel_stringstream << prefix_prevparalevel << ".sum"; + + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + INADM_string = INADM_stringstream.str(); const char* INADM_Cstr = INADM_string.c_str(); + CONV0_string = CONV0_stringstream.str(); const char* CONV0_Cstr = CONV0_string.c_str(); + STAT_string = STAT_stringstream.str(); const char* STAT_Cstr = STAT_string.c_str(); + LOG_string = LOG_stringstream.str(); const char* LOG_Cstr = LOG_string.c_str(); + THR_string = THR_stringstream.str(); const char* THR_Cstr = THR_string.c_str(); + SRC_string = SRC_stringstream.str(); const char* SRC_Cstr = SRC_string.c_str(); + //FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + SUM_string = SUM_stringstream.str(); const char* SUM_Cstr = SUM_string.c_str(); + //SUM_prevparalevel_string = SUM_prevparalevel_stringstream.str(); const char* SUM_prevparalevel_Cstr = SUM_prevparalevel_string.c_str(); + + THRDIR_string = THRDIR_stringstream.str(); + + fstream RAW_outfile; + if (!refine || in_parallel) RAW_outfile.open(RAW_Cstr, fstream::out | fstream::trunc); + else RAW_outfile.open(RAW_Cstr, fstream::out | fstream::app); + if (RAW_outfile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("Could not open RAW_outfile... "); + } + RAW_outfile.precision(16); + + fstream INADM_outfile; + if (!refine || in_parallel) INADM_outfile.open(INADM_Cstr, fstream::out | fstream::trunc); + else INADM_outfile.open(INADM_Cstr, fstream::out | fstream::app); + if (INADM_outfile.fail()) JSCerror("Could not open INADM_outfile... "); + INADM_outfile.precision(16); + + fstream CONV0_outfile; + if (!refine || in_parallel) CONV0_outfile.open(CONV0_Cstr, fstream::out | fstream::trunc); + else CONV0_outfile.open(CONV0_Cstr, fstream::out | fstream::app); + if (CONV0_outfile.fail()) JSCerror("Could not open CONV0_outfile... "); + CONV0_outfile.precision(16); + + fstream STAT_outfile; + if (!refine || in_parallel) STAT_outfile.open(STAT_Cstr, fstream::out | fstream::trunc); + else STAT_outfile.open(STAT_Cstr, fstream::out | fstream::app); + if (STAT_outfile.fail()) JSCerror("Could not open STAT_outfile... "); + STAT_outfile.precision(8); + + ofstream LOG_outfile; + if (!in_parallel) { + if (!refine) LOG_outfile.open(LOG_Cstr, fstream::out | fstream::trunc); + else LOG_outfile.open(LOG_Cstr, fstream::out | fstream::app); + if (LOG_outfile.fail()) JSCerror("Could not open LOG_outfile... "); + LOG_outfile.precision(16); + } + else { // in_parallel + LOG_outfile.open(LOG_Cstr, fstream::out | fstream::trunc); + if (LOG_outfile.fail()) JSCerror("Could not open LOG_outfile... "); + LOG_outfile.precision(16); + //LOG_outfile << endl; + } + + Scan_Info scan_info; + + //Scan_Thread_Set paused_thread_set; + //Scan_Thread_Set paused_thread_set_this_run; + if (!refine) mkdir(THRDIR_string.c_str(), S_IRWXU | S_IRWXG | S_IRWXO); + Scan_Thread_Data paused_thread_data (THRDIR_string, refine); + + if (refine) { + //paused_thread_set.Load(THR_Cstr); + paused_thread_data.Load(); + if (!in_parallel) scan_info.Load(SRC_Cstr); + } + + Scan_Info scan_info_before = scan_info; // for LOG file + Scan_Info scan_info_before_descent; + Scan_Info scan_info_obtained_in_descent; + + Scan_State_List ScanStateList (whichDSF, SeedScanState); + ScanStateList.Populate_List(whichDSF, SeedScanState); + + if (refine && !in_parallel) ScanStateList.Load_Info (SUM_Cstr); + else if (in_parallel && rank.sum() == 0) {}; // do nothing, keep the info in the higher .sum file! + + DP Chem_Pot = Chemical_Potential (AveragingState); + DP sumrule_factor = Sumrule_Factor (whichDSF, AveragingState, Chem_Pot, iKmin, iKmax); + + //clock_t stop_time_local = clock(); + + + // Now go for it ! + + bool at_least_one_new_flag_raised = false; + + int Ndata_previous_cycle = 0; + + int ninadm = 0; // number of inadmissible states for which we save some info in .inadm file. Save first 1000. + int nconv0 = 0; // number of unconverged states for which we save some info in .conv0 file. Save first 1000. + + double start_time_omp = omp_get_wtime(); + double current_time_omp = omp_get_wtime(); + +#pragma omp parallel + do { + + int omp_thread_nr = omp_get_thread_num(); + + if ((paused_thread_data.lowest_il_with_nthreads_neq_0 == paused_thread_data.nlists - 1) + && omp_thread_nr > 0) { + double start_time_wait = omp_get_wtime(); + //cout << "omp_thread " << omp_thread_nr << " sleeping for 5 seconds... " << endl; + //sleep(5); // give time to master omp_thread to populate threads + double stop_time_wait; + do { + for (int i = 0; i < 100000; ++i) { } + stop_time_wait = omp_get_wtime(); + } while (stop_time_wait - start_time_wait < 5.0); + //cout << "omp_thread " << omp_thread_nr << " restarting" << endl; + } + + double start_time_cycle_omp = omp_get_wtime(); + + at_least_one_new_flag_raised = false; + + //if (!in_parallel) { // flag raising not allowed in parallel mode + //if (!in_parallel || rank == 0) { // flag raising only allowed if not in parallel mode, or if rank == 0 + //if (!in_parallel || rank.sum() == 0) { // flag raising only allowed if not in parallel mode, or if rank == 0 at all paralevels + + //if (!in_parallel) { // flag raising not allowed in parallel mode +#pragma omp master + { + //clock_t start_time_flags = clock(); + double start_time_flags = omp_get_wtime(); + + // First flag the new base/type 's that we need to include: + //thresholdremoved ScanStateList.Raise_Scanning_Flags (running_scan_threshold); + ScanStateList.Raise_Scanning_Flags (exp(-paused_thread_data.logscale * paused_thread_data.lowest_il_with_nthreads_neq_0)); + + //cout << "flags: " << endl << ScanStateList.flag_for_scan << endl; + + + // Get these base/type started: + for (int i = 0; i < ScanStateList.ndef; ++i) { + if (ScanStateList.flag_for_scan[i] && ScanStateList.info[i].Ndata == 0 && !ScanStateList.scan_attempted[i]) { + + Scan_Info scan_info_flags; + + at_least_one_new_flag_raised = true; + + ScanStateList.scan_attempted[i] = true; + + Tstate ScanState; + + //scan_info_before_descent = scan_info; + + ScanState = ScanStateList.State[i]; + + //cout << "ScanStateList.State[i] = " << ScanState << endl; + + DP data_value = -1.0; + + bool admissible = ScanState.Check_Admissibility(whichDSF); + + if (admissible) { + + ScanState.Compute_All(true); + + if (ScanState.conv) { + + // Put momentum in fundamental window, if possible: + int iKexc = ScanState.iK - AveragingState.iK; + while (iKexc > iKmax && iKexc - iKmod >= iKmin) iKexc -= iKmod; + while (iKexc < iKmin && iKexc + iKmod <= iKmax) iKexc += iKmod; + + //if (iKexc >= iKmin && iKexc <= iKmax) RAW_outfile << endl; + + //data_value = Compute_Matrix_Element_Contrib (whichDSF, iKmin, iKmax, ScanState, AveragingState, Chem_Pot, RAW_outfile); + stringstream rawfile_entry; + data_value = Compute_Matrix_Element_Contrib (whichDSF, iKmin, iKmax, ScanState, AveragingState, Chem_Pot, rawfile_entry); + { +#pragma omp critical + RAW_outfile << rawfile_entry.str(); + } + //cout << "data_value for ScanState.label " << ScanState.label << " = " << data_value << endl; + { +#pragma omp critical + if (iKexc >= iKmin && iKexc <= iKmax) { + scan_info_flags.Ndata++; + scan_info_flags.Ndata_conv++; + scan_info_flags.sumrule_obtained += data_value*sumrule_factor; + } + } + //cout << data_value * sumrule_factor << endl; + + // If we force descent: modify data_value by hand so that descent is forced on next scanning pass + //if (Force_Descent(whichDSF, ScanState, AveragingState, iKmod, Chem_Pot) && ScanState.iK - AveragingState.iK < iKmax && Sca nState.iK - AveragingState.iK > iKmin) + //if (Force_Descent(whichDSF, ScanState, AveragingState, iKmod, Chem_Pot)) + for (int itype = 0; itype < 15; ++itype) { + DP data_value_used = 0.1* exp(-paused_thread_data.logscale * JSC::min(0, paused_thread_data.lowest_il_with_nthreads_neq_0)); + if (Force_Descent(whichDSF, ScanState, AveragingState, itype, iKmod, Chem_Pot)) + //data_value = 0.1* exp(-paused_thread_data.logscale * paused_thread_data.lowest_il_with_nthreads_neq_0); + data_value = data_value_used; + } + // ++G_7 logic + Vect allowed(false, 15); + if (whichDSF == 'B') { // symmetric state scanning + allowed[9] = true; allowed[10] = true; allowed[11] = true; + allowed[12] = true; allowed[13] = true; allowed[14] = true; + } + else { + allowed[0] = (iKexc >= iKmin && iKexc <= iKmax); + allowed[1] = allowed[0]; allowed[2] = allowed[0]; allowed[3] = allowed[0]; allowed[4] = allowed[0]; allowed[5] = allowed[0]; allowed[6] = allowed[0]; allowed[7] = allowed[0]; allowed[8] = allowed[0]; + //allowed[9] = (iKexc <= 0 && iKexc > iKmin); + allowed[9] = (iKexc > iKmin); + allowed[10] = allowed[9]; allowed[11] = allowed[9]; + //allowed[12] = (iKexc >= 0 && iKexc < iKmax); + allowed[12] = (iKexc < iKmax); + allowed[13] = allowed[12]; allowed[14] = allowed[12]; + } + for (int type_required_here = 0; type_required_here < 15; ++type_required_here) { + if (!allowed[type_required_here]) continue; + // All cases here are such that the ScanState hasn't been descended yet, so we simply use data_value as expected data value: + //paused_thread_set_this_run.Include_Thread (abs(data_value), ScanState.label, type_required_here); + { +#pragma omp critical + paused_thread_data.Include_Thread (abs(data_value), ScanState.label, type_required_here); + } + } + } // if (ScanState.conv) + + else { + if (nconv0++ < 1000) + CONV0_outfile << setw(25) << ScanState.label << setw(25) << ScanState.diffsq << setw(5) << ScanState.Check_Rapidities() + << setw(25) << ScanState.String_delta() << endl; + scan_info_flags.Ndata++; + scan_info_flags.Ndata_conv0++; + } + } // if admissible + + else { // if inadmissible, modify data_value by hand so that descent is forced on next scanning pass + + if (ninadm++ < 10000000) INADM_outfile << ScanState.label << endl; + scan_info_flags.Ndata++; + scan_info_flags.Ninadm++; + + // Put momentum in fundamental window, if possible: + int iKexc = ScanState.iK - AveragingState.iK; + while (iKexc > iKmax && iKexc - iKmod >= iKmin) iKexc -= iKmod; + while (iKexc < iKmin && iKexc + iKmod <= iKmax) iKexc += iKmod; + + //thresholdremoved DP data_value = 2.0* running_scan_threshold; + //DP data_value = 2.0 * exp(-paused_thread_data.logscale * paused_thread_data.lowest_il_with_nthreads_neq_0); + //DP data_value = 0.1* running_scan_threshold; + DP data_value = 1.0e-32; + for (int itype = 0; itype < 15; ++itype) + if (Force_Descent(whichDSF, ScanState, AveragingState, itype, iKmod, Chem_Pot)) + data_value = 0.1* exp(-paused_thread_data.logscale * paused_thread_data.lowest_il_with_nthreads_neq_0); + + // ++G_7 logic + Vect allowed(false, 15); + if (whichDSF == 'B') { + // We scan over symmetric states. Only types 14 down to 9 are allowed. + allowed[9] = true; allowed[10] = true; allowed[11] = true; + allowed[12] = true; allowed[13] = true; allowed[14] = true; + } + else { + allowed[0] = (iKexc >= iKmin && iKexc <= iKmax); + allowed[1] = allowed[0]; allowed[2] = allowed[0]; allowed[3] = allowed[0]; allowed[4] = allowed[0]; allowed[5] = allowed[0]; allowed[6] = allowed[0]; allowed[7] = allowed[0]; allowed[8] = allowed[0]; + //allowed[9] = (iKexc <= 0 && iKexc > iKmin); + allowed[9] = (iKexc > iKmin); + allowed[10] = allowed[9]; allowed[11] = allowed[9]; + //allowed[12] = (iKexc >= 0 && iKexc < iKmax); + allowed[12] = (iKexc < iKmax); + allowed[13] = allowed[12]; allowed[14] = allowed[12]; + } + for (int type_required_here = 0; type_required_here < 15; ++type_required_here) { + if (!allowed[type_required_here]) continue; + //paused_thread_set_this_run.Include_Thread (abs(data_value), ScanState.label, type_required_here); + { +#pragma omp critical + paused_thread_data.Include_Thread (abs(data_value), ScanState.label, type_required_here); + } + } + } // inadmissible + + //scan_info_obtained_in_descent = scan_info; + //scan_info_obtained_in_descent -= scan_info_before_descent; + + scan_info_flags.TT += omp_get_wtime() - start_time_flags; + + // Put this info into the appropriate ScanStateList.info + { +#pragma omp critical + //ScanStateList.Include_Info(scan_info_obtained_in_descent, ScanStateList.base_label[i]); + //cout << "Including info_flags: " << scan_info_flags << endl; + ScanStateList.Include_Info(scan_info_flags, ScanStateList.base_label[i]); + scan_info += scan_info_flags; + } + + //cout << "Done with state " << ScanState.label << endl; + + } // if flag_for_scan + } // for i + + //clock_t stop_time_flags = clock(); + + //scan_info.CPU_ticks += stop_time_flags - start_time_flags; + //scan_info.TT += (stop_time_flags - start_time_flags)/CLOCKS_PER_SEC; + //scan_info.TT += omp_get_wtime() - start_time_flags; + + //} // if (!in_parallel || rank == 0) + + } // #pragma omp master + + //cout << "Done raising flags." << endl; + + // Now we deal with the previously existing paused threads: + + //if (scan_info.CPU_ticks < ((long long int) Max_Secs) * ((long long int) CLOCKS_PER_SEC)) { + + //Vect threads_to_do = paused_thread_data.Extract_Next_Scan_Threads(); + Vect threads_to_do; + int il_to_do = paused_thread_data.lowest_il_with_nthreads_neq_0; // for resaving threads in case we're out of time + { +#pragma omp critical + threads_to_do = paused_thread_data.Extract_Next_Scan_Threads(); + //threads_to_do = paused_thread_data.Extract_Next_Scan_Threads(100); + } + //cout << "Size of threads_to_do: " << threads_to_do.size() << endl; + //for (int i = 0; i < threads_to_do.size(); ++i) cout << threads_to_do[i].label << "\t" << threads_to_do[i].type << "\t"; + //cout << endl; + + int ithread; + + //omp1#pragma omp parallel + { + //omp1#pragma omp for + for (ithread = 0; ithread < threads_to_do.size(); ++ithread) { + + //cout << "\tithread = " << ithread << endl; + //scan_info_before_descent = scan_info; + + //int tid = omp_get_thread_num(); + //{ + //#pragma omp critical + //cout << "Thread " << tid << " handling ithread " << ithread << " out of " << threads_to_do.size() << "\t" << threads_to_do[ithread].label << "\t" << threads_to_do[ithread].type << endl; + //} + + Scan_Info scan_info_this_ithread; + double start_time_this_ithread = omp_get_wtime(); + + // If we don't have time anymore, resave the threads instead of computing them: + if (start_time_this_ithread - start_time_omp > Max_Secs_alert) { + for (int ith = ithread; ith < threads_to_do.size(); ++ith) { +#pragma omp critical + paused_thread_data.Include_Thread (il_to_do, threads_to_do[ith].label, threads_to_do[ith].type); + } + break; // jump out of ithread loop + } + + Tstate ScanState; + { +#pragma omp critical + ScanState = ScanStateList.Return_State(Extract_Base_Label(threads_to_do[ithread].label)); + } + Tstate BaseScanState; BaseScanState = ScanState; + + //cout << "Setting to label = " << threads_to_do[ithread].label << ", descending type " << threads_to_do[ithread].type << endl; + ScanState.Set_to_Label(threads_to_do[ithread].label, BaseScanState.Ix2); + //cout << "ScanState after setting label: " << threads_to_do[ithread].label << endl << ScanState << endl; + + + //cout << "Calling Descend_and_Compute with type " << paused_thread_list.type[ithread] << " on state" << endl << ScanState << endl; + /* + Descend_and_Compute_for_Fixed_Base (whichDSF, AveragingState, BaseScanState, ScanState, + //paused_thread_set.type[ilist][ithread], iKmin, iKmax, iKmod, + threads_to_do[ithread].type, iKmin, iKmax, iKmod, + //paused_thread_set_this_run, running_scan_threshold, + paused_thread_data, //thresholdremoved running_scan_threshold, + //paused_thread_set[ilist].abs_data_value[ithread], + ph_cost, + Max_Secs_used, sumrule_factor, Chem_Pot, scan_info, RAW_outfile, + INADM_outfile, ninadm, CONV0_outfile, nconv0, STAT_outfile); + */ + + // STARTING Descend_and_Compute block: + int type_required = threads_to_do[ithread].type; + + ScanState.Compute_Momentum(); + Vect desc_label; + + // ++G_7 logic + bool disperse_only_current_exc_up = false; + if (type_required == 14 || type_required == 8 || type_required == 7 || type_required == 6) disperse_only_current_exc_up = true; + bool preserve_nexc_up = false; + if (type_required == 13 || type_required == 5 || type_required == 4 || type_required == 3) preserve_nexc_up = true; + bool disperse_only_current_exc_down = false; + if (type_required == 11 || type_required == 8 || type_required == 5 || type_required == 2) disperse_only_current_exc_down = true; + bool preserve_nexc_down = false; + if (type_required == 10 || type_required == 7 || type_required == 4 || type_required == 1) preserve_nexc_down = true; + + if (whichDSF == 'B') { // symmetric state scanning + if (type_required >= 9 && type_required <= 11) + desc_label = Descendent_States_with_iK_Stepped_Down_rightIx2only (ScanState.label, BaseScanState, disperse_only_current_exc_down, preserve_nexc_down); + else if (type_required >= 12 && type_required <= 14) + desc_label = Descendent_States_with_iK_Stepped_Up_rightIx2only (ScanState.label, BaseScanState, disperse_only_current_exc_up, preserve_nexc_up); + } + else { + if (type_required >= 0 && type_required <= 8) { + desc_label = Descendent_States_with_iK_Preserved(ScanState.label, BaseScanState, disperse_only_current_exc_up, preserve_nexc_up, disperse_only_current_exc_down, preserve_nexc_down); + } + else if (type_required >= 9 && type_required <= 11) + desc_label = Descendent_States_with_iK_Stepped_Down (ScanState.label, BaseScanState, disperse_only_current_exc_down, preserve_nexc_down); + else if (type_required >= 12 && type_required <= 14) + desc_label = Descendent_States_with_iK_Stepped_Up (ScanState.label, BaseScanState, disperse_only_current_exc_up, preserve_nexc_up); + } + + string label_here = ScanState.label; + //int ScanState_iK = ScanState.iK; + + for (int idesc = 0; idesc < desc_label.size(); ++idesc) { + + //clock_t start_time_here = clock(); + + ScanState.Set_to_Label (desc_label[idesc], BaseScanState.Ix2); + + bool admissible = ScanState.Check_Admissibility(whichDSF); + + DP data_value = 0.0; + + //scan_info.Ndata++; + //scan_info_this_ithread.Ndata++; + + ScanState.conv = false; + ScanState.Compute_Momentum(); // since momentum is used as forced descent criterion + + + if (admissible) { + + ScanState.Compute_All (idesc == 0); + //ScanState.Compute_All (true); + + //scan_info.Ndata++; + + if (ScanState.conv) { + //scan_info_this_ithread.Ndata_conv++; + + // Put momentum in fundamental window, if possible: + int iKexc = ScanState.iK - AveragingState.iK; + while (iKexc > iKmax && iKexc - iKmod >= iKmin) iKexc -= iKmod; + while (iKexc < iKmin && iKexc + iKmod <= iKmax) iKexc += iKmod; + + //data_value = Compute_Matrix_Element_Contrib (whichDSF, iKmin, iKmax, ScanState, AveragingState, Chem_Pot, RAW_outfile); + stringstream rawfile_entry; + data_value = Compute_Matrix_Element_Contrib (whichDSF, iKmin, iKmax, ScanState, AveragingState, Chem_Pot, rawfile_entry); + + { +#pragma omp critical + RAW_outfile << rawfile_entry.str(); + if (iKexc >= iKmin && iKexc <= iKmax) { + scan_info_this_ithread.Ndata++; + scan_info_this_ithread.Ndata_conv++; + scan_info_this_ithread.sumrule_obtained += data_value*sumrule_factor; + } + } + + //if (iKexc >= iKmin && iKexc <= iKmax) scan_info_this_ithread.sumrule_obtained += data_value*sumrule_factor; + //cout << "data_value found = " << data_value * sumrule_factor << endl; + + // Uncomment line below if .stat file is desired: + //STAT_outfile << setw(20) << label_here << "\t" << setw(5) << type_required << "\t" << setw(16) << std::scientific << running_scan_threshold << "\t" << setw(20) << ScanState.label << "\t" << setw(16) << data_value << "\t" << setw(16) << std::fixed << setprecision(8) << data_value/running_scan_threshold << endl; + + } // if (ScanState.conv) + else { + if (nconv0++ < 1000) + CONV0_outfile << setw(25) << ScanState.label << setw(25) << ScanState.diffsq << setw(5) << ScanState.Check_Rapidities() + << setw(25) << ScanState.String_delta() << endl; + //scan_info.Ndata_conv0++; + scan_info_this_ithread.Ndata++; + scan_info_this_ithread.Ndata_conv0++; + //cout << "State did not converge." << endl; + } + } // if (admissible) + + else { + if (ninadm++ < 1000000) INADM_outfile << ScanState.label << endl; + //scan_info.Ninadm++; + scan_info_this_ithread.Ndata++; + scan_info_this_ithread.Ninadm++; + //cout << "State was inadmissible." << endl; + // Set data_value to enable continued scanning later on: + //thresholdremoved data_value = 0.1* running_scan_threshold; + } + + + //clock_t stop_time_here = clock(); + + //scan_info.CPU_ticks += stop_time_here - start_time_here; + //scan_info_this_ithread.CPU_ticks += stop_time_here - start_time_here; + //scan_info_this_ithread.TT += (stop_time_here - start_time_here)/CLOCKS_PER_SEC; + + Tstate state_to_descend; state_to_descend = ScanState; // for checking + + ScanState.Compute_Momentum(); + // Put momentum in fundamental window, if possible: + int iKexc = ScanState.iK - AveragingState.iK; + while (iKexc > iKmax && iKexc - iKmod >= iKmin) iKexc -= iKmod; + while (iKexc < iKmin && iKexc + iKmod <= iKmax) iKexc += iKmod; + + // ++G_7 logic + // Momentum-preserving are only descended to momentum-preserving. + // Momentum-increasing are only descended to momentum-preserving and momentum-increasing. + // Momentum-decreasing are only descended to momentum-preserving and momentum-decreasing. + Vect allowed(false, 15); + if (whichDSF == 'B') { + // We scan over symmetric states. Only types 14 down to 9 are allowed. + if (type_required >= 9 && type_required <= 11) { // iK stepped down on rightIx2; step further up or down + allowed[9] = true; allowed[10] = true; allowed[11] = true; + allowed[12] = true; allowed[13] = true; allowed[14] = true; + } + else if (type_required >= 12 && type_required <= 14) { // iK stepped up on rightIx2; only step further up + allowed[12] = true; allowed[13] = true; allowed[14] = true; + } + } + else { + if (type_required >= 0 && type_required <= 8) { // momentum-preserving + allowed[0] = (iKexc >= iKmin && iKexc <= iKmax); + allowed[9] = false; + allowed[12] = false; + } + if (type_required >= 9 && type_required <= 11) { // momentum-decreasing + allowed[0] = (iKexc >= iKmin && iKexc <= iKmax); + allowed[9] = (iKexc > iKmin); + allowed[12] = false; + } + if (type_required >= 12 && type_required <= 14) { // momentum-increasing + allowed[0] = (iKexc >= iKmin && iKexc <= iKmax); + allowed[9] = false; + allowed[12] = (iKexc < iKmax); + } + // The others are just copies of the ones above: + allowed[1] = allowed[0]; allowed[2] = allowed[0]; allowed[3] = allowed[0]; allowed[4] = allowed[0]; allowed[5] = allowed[0]; allowed[6] = allowed[0]; allowed[7] = allowed[0]; allowed[8] = allowed[0]; + allowed[10] = allowed[9]; allowed[11] = allowed[9]; + allowed[13] = allowed[12]; allowed[14] = allowed[12]; + } + + + for (int type_required_here = 0; type_required_here < 15; ++type_required_here) { + + if (!allowed[type_required_here]) continue; + + // Reset ScanState to what it was, if change on first pass + if (type_required_here > 0) ScanState = state_to_descend; + + // We determine if we carry on scanning based on the data_value obtained, or forcing conditions: + // Forcing conditions: + //if (!admissible || Force_Descent(whichDSF, ScanState, AveragingState, type_required_here, iKmod, Chem_Pot)) + ////data_value = 1.01 * running_scan_threshold/ph_cost; // force for all types of desc + //data_value = 1.01 * running_scan_threshold; // only force for no new ph pairs + ////data_value = 1.0; // force for all types of desc + // If we're sitting out of the iKmin & iKmax window, stop: + //if (iKmin != iKmax && (ScanState.iK - AveragingState.iK > iKmax || ScanState.iK - AveragingState.iK < iKmin)) data_value = 0.0; + + //if (abs(data_value) * (type_required_here != 2 ? 1.0 : ph_cost) > running_scan_threshold + //if ((abs(data_value) > running_scan_threshold + //|| Nr_ph_Recombinations_Possible (ScanState.label, BaseScanState, type_required_here) > 0) + + //DP expected_abs_data_value = abs(data_value)/pow(ph_cost, DP(Nr_ph_Recombinations_Possible (ScanState.label, BaseScanState, type_required_here))); + DP expected_abs_data_value = abs(data_value); + + //++G_7 logic + if ((type_required_here == 14 || type_required_here == 8 || type_required_here == 7 || type_required_here == 6) + && Expect_ph_Recombination_iK_Up (ScanState.label, BaseScanState)) expected_abs_data_value /= ph_cost; + if (type_required_here == 12 || type_required_here == 2 || type_required_here == 1 || type_required_here == 0) + expected_abs_data_value *= ph_cost; + if ((type_required_here == 11 || type_required_here == 8 || type_required_here == 5 || type_required_here == 2) + && Expect_ph_Recombination_iK_Down (ScanState.label, BaseScanState)) expected_abs_data_value /= ph_cost; + if (type_required_here == 9 || type_required_here == 6 || type_required_here == 3 || type_required_here == 0) + expected_abs_data_value *= ph_cost; + + //paused_thread_set.Include_Thread (expected_abs_data_value, ScanState.label, type_required_here); + { +#pragma omp critical + //cout << "\tIncluding thread " << ScanState.label << "\t" << type_required_here << "\tdata_value " << data_value << "\texpected abs data value " << expected_abs_data_value << endl; + paused_thread_data.Include_Thread (expected_abs_data_value, ScanState.label, type_required_here); + } + //cout << "\tDone including thread." << endl; + } // for type_required_here + + //cout << "\tFinished with descendent " << idesc << " out of " << desc_label.size() << " with label " << desc_label[idesc] << endl; + //cout << "\tfrom state with label " << label_here << endl; + } // for idesc + + //cout << "Finished Descend on state " << endl << ScanState.label << endl; + + // FINISHED Descend_and_Compute block + + + + //cout << "Finished descending." << endl; + //cout << "\tFinished descending ithread = " << ithread << endl; + + //scan_info_obtained_in_descent = scan_info; + //scan_info_obtained_in_descent -= scan_info_before_descent; + //// Put this info into the appropriate ScanStateList.info + //ScanStateList.Include_Info(scan_info_obtained_in_descent, Extract_Base_Label(threads_to_do[ithread].label)); + + scan_info_this_ithread.TT += omp_get_wtime() - start_time_this_ithread; + +#pragma omp critical + { + scan_info += scan_info_this_ithread; + //cout << "Including info_ihtread: " << scan_info_this_ithread << endl; + ScanStateList.Include_Info(scan_info_this_ithread, Extract_Base_Label(threads_to_do[ithread].label)); + } + } // for ithread + + } // omp parallel region + + // Resynchronize all compute threads: + //omp1#pragma omp barrier + + //cout << "Done with threads_to_do." << endl; + + // } // if time + + //start_time_local = clock(); + + /* + if (!in_parallel) + LOG_outfile << "Ndata handled up to now: " << scan_info.Ndata_conv + << ". Threshold level " << paused_thread_data.lowest_il_with_nthreads_neq_0 << " " << setprecision(6) << exp(-paused_thread_data.logscale * paused_thread_data.lowest_il_with_nthreads_neq_0) + << ". " << scan_info.Ndata - Ndata_previous_cycle << " new data points. Number of threads: " + << paused_thread_data.nthreads_total.sum() + << ". Saturation: " << setprecision(12) << scan_info.sumrule_obtained << endl; + */ + + //int tid = omp_get_thread_num(); +#pragma omp master + { + if (!in_parallel) + LOG_outfile << "Master cycling. Ndata_conv " << scan_info.Ndata_conv + << ". Threshold " << paused_thread_data.lowest_il_with_nthreads_neq_0 << " " << setw(9) << setprecision(3) << exp(-paused_thread_data.logscale * paused_thread_data.lowest_il_with_nthreads_neq_0) + << ". " << setw(12) << scan_info.Ndata - Ndata_previous_cycle << " new data. Nr of threads: " + << setw(14) << paused_thread_data.nthreads_total.sum() + << ". Saturation: " << setprecision(12) << scan_info.sumrule_obtained << endl; + + Ndata_previous_cycle = scan_info.Ndata; + } + + + //stop_time_local = clock(); + + //current_time = clock(); + + //scan_info.CPU_ticks += stop_time_local - start_time_local; + + //#pragma omp critical + //scan_info.TT += omp_get_wtime() - start_time_cycle_omp; + + current_time_omp = omp_get_wtime(); + + //int tid = omp_get_thread_num(); + //if (tid == 0) cout << "current_time_omp - start_time_omp = " << current_time_omp - start_time_omp << "\t" << Max_Secs_used << endl; + //if (current_time_omp - start_time_omp > Max_Secs_used) + //cout << "tid " << tid << " exiting." << endl; + + + //} while (scan_info.CPU_ticks < ((long long int) Max_Secs_used) * ((long long int) CLOCKS_PER_SEC) + //} while (current_time - start_time < ((long long int) Max_Secs_used) * ((long long int) CLOCKS_PER_SEC) + } while (current_time_omp - start_time_omp < Max_Secs_used + && scan_info.sumrule_obtained < target_sumrule + //&& paused_thread_list.Highest_abs_data_value(0.0, 1.0e+10) > 1.0e-30 + //&& !(all_threads_zero_previous_cycle && all_threads_zero_current_cycle && !at_least_one_new_flag_raised) + //thresholdremoved && running_scan_threshold > 1.0e-10*MACHINE_EPS + ); + // This closes the #pragram omp parallel block + + RAW_outfile.close(); + INADM_outfile.close(); + CONV0_outfile.close(); + STAT_outfile.close(); + + + //scan_info.CPU_ticks_TOT += scan_info.CPU_ticks; + scan_info.Save(SRC_Cstr); + + Scan_Info scan_info_refine = scan_info; + scan_info_refine -= scan_info_before; + + if (!in_parallel) { + + if (scan_info.sumrule_obtained >= target_sumrule) + LOG_outfile << endl << "Achieved sumrule saturation of " << scan_info.sumrule_obtained + << "\t(target was " << target_sumrule << ")." << endl << endl; + + //thresholdremoved + //if (running_scan_threshold < MACHINE_EPS) + //LOG_outfile << endl << "Stopping because threshold lower than machine precision. " << endl << endl; + + //thresholdremoved if (!refine) LOG_outfile << "Main run info: " << scan_info << endl << "Latest running_scan_threshold = " << running_scan_threshold << endl; + if (!refine) { + LOG_outfile << "Main run info: " << scan_info << endl; + LOG_outfile << "Latest threshold level " << paused_thread_data.lowest_il_with_nthreads_neq_0 << " " << std::scientific << setprecision(3) << exp(-paused_thread_data.logscale * paused_thread_data.lowest_il_with_nthreads_neq_0) << endl; + } + else if (refine) { + //thresholdremoved LOG_outfile << "Refining info: " << scan_info_refine << endl << "Latest running_scan_threshold = " << running_scan_threshold << endl + LOG_outfile << "Refining info: " << scan_info_refine << endl; + LOG_outfile << "Latest threshold level " << paused_thread_data.lowest_il_with_nthreads_neq_0 << " " << std::scientific << setprecision(3) << exp(-paused_thread_data.logscale * paused_thread_data.lowest_il_with_nthreads_neq_0) << endl; + LOG_outfile << "Resulting info: " << scan_info << endl; + } + LOG_outfile << "Code version " << JSC_VERSION << ", copyright J.-S. Caux." << endl << endl; + LOG_outfile.close(); + } + + else { // in_parallel + + //thresholdremoved + //if (running_scan_threshold < MACHINE_EPS) + //LOG_outfile << "rank " << rank << " out of " << nr_processors << " processors: " + // << "Stopping because threshold lower than machine precision. " << endl << endl; + + LOG_outfile << "rank " << rank << " out of " << nr_processors << " processors: " + //thresholdremoved << "run info: " << scan_info << endl << "Latest running_scan_threshold = " << running_scan_threshold << endl; + << "run info: " << scan_info << endl << "Latest threshold = " << exp(-paused_thread_data.logscale * paused_thread_data.lowest_il_with_nthreads_neq_0) << endl; + } + + //if (paused_thread_list.dim < 1000000) paused_thread_list.Order_in_abs_data_value(); + //paused_thread_list.Save(THR_Cstr); + //paused_thread_set.Save(THR_Cstr); + paused_thread_data.Save(); + + ScanStateList.Order_in_SRC (); + + //cout << "Saving info: " << endl; for (int idef = 0; idef < ScanStateList.ndef; ++idef) cout << ScanStateList.info[idef] << endl; + ScanStateList.Save_Info (SUM_Cstr); + + + // Evaluate f-sumrule: + //if (!fixed_iK && !in_parallel) if (whichDSF != 'q') Evaluate_F_Sumrule (whichDSF, AveragingState, Chem_Pot, RAW_Cstr, FSR_Cstr); + //if (iKmin != iKmax && !in_parallel) if (whichDSF != 'q') Evaluate_F_Sumrule (whichDSF, AveragingState, Chem_Pot, iKmin, iKmax, RAW_Cstr, FSR_Cstr); + //if (iKmin != iKmax && !in_parallel ) if (whichDSF != 'q') Evaluate_F_Sumrule (prefix_prevparalevel, whichDSF, AveragingState, Chem_Pot, iKmin, iKmax); + if (!in_parallel ) if (whichDSF != 'q') Evaluate_F_Sumrule (prefix_prevparalevel, whichDSF, AveragingState, Chem_Pot, iKmin, iKmax); + + // Produce sorted file + //if (!in_parallel && whichDSF != 'Z') Sort_RAW_File (RAW_Cstr, 'f', whichDSF); + //if (!in_parallel && whichDSF == 'Z') Sort_RAW_File (RAW_Cstr, 'e', whichDSF); + + return(scan_info); + + + } + + + + + //****************************************************** + + // Functions to initiate scans: + + + // General version for equilibrium correlators at generic (possibly finite) temperature: + + void Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors) + { + + // This function scans the Hilbert space of the LiebLin gas, + // for the function identified by whichDSF. + + // whichDSF == 'Z': canonical partition function + // whichDSF == 'd': density-density correlation function + // whichDSF == 'g': Green's function < \Psi \Psi^{\dagger}> + // whichDSF == 'o': one-body function < \Psi^{\dagger} \Psi > + + // Delta is the number of sites involved in the smoothing of the entropy + //int Delta = int(sqrt(N))/2;//6;//N/20; + //DP epsilon = log(L)/L; // using Gaussian for density in entropy. + //DP epsilon = 1.0/L; // using Lorentzian for density in entropy. + + // Construct the finite-size saddle-point state: + // if we refine, read the quantum numbers of the saddle point state (and seed sps) from the sps file: + + stringstream SPS_stringstream; string SPS_string; + //SPS_stringstream << "Tgt0_"; + //Data_File_Name (SPS_stringstream, whichDSF, iKmin, iKmax, kBT, spstate, SeedScanState, ""); + Data_File_Name (SPS_stringstream, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + SPS_stringstream << ".sps"; + SPS_string = SPS_stringstream.str(); const char* SPS_Cstr = SPS_string.c_str(); + + fstream spsfile; + if (refine) spsfile.open(SPS_Cstr, fstream::in); + else spsfile.open(SPS_Cstr, fstream::out | fstream::trunc); + if (spsfile.fail()) { + cout << SPS_Cstr << endl; JSCerror("Could not open spsfile."); + } + + LiebLin_Bethe_State spstate; + + if (!refine) { // obtain the sps from discretized TBA + spstate = Canonical_Saddle_Point_State (c_int, L, N, whichDSF == 'Z' ? 0.0 : kBT); + } + else { // read it from the sps file + // Check that the sps has the right number of Ix2: + int Nspsread; + spsfile >> Nspsread; + if (Nspsread != N) { + cout << Nspsread << "\t" << N << endl; + JSCerror("Wrong number of Ix2 in saddle-point state."); + } + spstate = LiebLin_Bethe_State (c_int, L, N); + for (int i = 0; i < N; ++i) spsfile >> spstate.Ix2[i]; + } + spstate.Compute_All(true); + + int Nscan = N; + if (whichDSF == 'o') Nscan = N - 1; + if (whichDSF == 'g') Nscan = N + 1; + + // Now construct or read off the seed scan state: + + // TO MODIFY: this is not a good idea, since this might construct a state with many p-h w/r to the AveragingState. + LiebLin_Bethe_State SeedScanState; + + if (whichDSF != 'o' && whichDSF != 'g') SeedScanState = spstate; + + else if (whichDSF == 'o' || whichDSF == 'g') { + if (!refine) { + //SeedScanState = Canonical_Saddle_Point_State (c_int, L, Nscan, kBT); + //LiebLin_Bethe_State scanspstate = Canonical_Saddle_Point_State (c_int, L, Nscan, kBT); + //SeedScanState = scanspstate; + if (whichDSF == 'o') SeedScanState = Remove_Particle_at_Center (spstate); + else SeedScanState = Add_Particle_at_Center (spstate); + } + else { // read it from the sps file + // Check that the sps has the right number of Ix2: + int Nsspsread; + spsfile >> Nsspsread; + if (Nsspsread != Nscan) { + cout << Nsspsread << "\t" << Nscan << endl; + JSCerror("Wrong number of Ix2 in scan saddle-point state."); + } + SeedScanState = LiebLin_Bethe_State (c_int, L, Nscan); + for (int i = 0; i < Nscan; ++i) spsfile >> SeedScanState.Ix2[i]; + } + } // if one-body or Green's function + + SeedScanState.Compute_All(true); + + LiebLin_Bethe_State ScanState = SeedScanState; + + DP delta = sqrt(DP(N)) * (spstate.lambdaoc[N-1] - spstate.lambdaoc[0])/N; + + if (!refine) { // we write data to the sps file + + spsfile << N << endl; + spsfile << spstate.Ix2 << endl; + spsfile << Nscan << endl; + spsfile << SeedScanState.Ix2 << endl; + + spsfile << endl << spstate << endl << endl; + for (int i = 1; i < spstate.N - 2; ++i) + spsfile << 0.5 * (spstate.lambdaoc[i] + spstate.lambdaoc[i+1]) + //<< "\t" << twoPI/(spstate.L * (spstate.lambdaoc[i+1] - spstate.lambdaoc[i])) << endl; + << "\t" << 1.0/spstate.L * (0.25/(spstate.lambdaoc[i] - spstate.lambdaoc[i-1]) + + 0.5/(spstate.lambdaoc[i+1] - spstate.lambdaoc[i]) + + 0.25/(spstate.lambdaoc[i+2] - spstate.lambdaoc[i+1])) + << "\t" << rho_of_lambdaoc_1 (spstate, 0.5 * (spstate.lambdaoc[i] + spstate.lambdaoc[i+1]), delta) + << "\t" << rho_of_lambdaoc_2 (spstate, 0.5 * (spstate.lambdaoc[i] + spstate.lambdaoc[i+1]), delta) + << endl; + } + spsfile.close(); + + // Perform the scan: + General_Scan (whichDSF, iKmin, iKmax, 100000000, kBT, spstate, SeedScanState, "", Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + return; + } + + void Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, + int Max_Secs, DP target_sumrule, bool refine) + { + int paralevel = 0; + Vect rank(0,1); + Vect nr_processors(0,1); + + Scan_LiebLin (whichDSF, c_int, L, N, iKmin, iKmax, kBT, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + return; + } + + + // Scanning on an excited state defined by a set of Ix2: + void Scan_LiebLin (char whichDSF, LiebLin_Bethe_State AveragingState, string defaultScanStatename, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors) + { + // This function is as Scan_LiebLin for generic T defined above, except that the + // averaging is now done on a state defined by AveragingStateIx2 + + // PRECONDITIONS: + // - the Ix2 of AveragingState are properly set. + + DP c_int = AveragingState.c_int; + DP L = AveragingState.L; + int N = AveragingState.N; + + //LiebLin_Bethe_State GroundState (c_int, L, N); + // Make sure the label of AveragingState is properly set to that relative to GS: + //AveragingState.Set_Label_from_Ix2 (GroundState.Ix2); + + // The label of the Averaging State is by definition the `empty' label + AveragingState.Set_Label_from_Ix2 (AveragingState.Ix2); + AveragingState.Compute_All(true); + + int Nscan = N; + if (whichDSF == 'o') Nscan = N - 1; + if (whichDSF == 'g') Nscan = N + 1; + + LiebLin_Bethe_State SeedScanState (c_int, L, Nscan); + if (whichDSF == 'd' || whichDSF == 'B') SeedScanState.Ix2 = AveragingState.Ix2; + // If 'o', remove rightmost and shift quantum numbers by half-integer towards center. + // if (whichDSF == 'o') for (int i = 0; i < N-1; ++i) SeedScanState.Ix2[i] = AveragingState.Ix2[i] + 1; + // If 'g', add a new particle at the right, after shifting all towards center. + //if (whichDSF == 'g') { + //for (int i = 0; i < N; ++i) SeedScanState.Ix2[i] = AveragingState.Ix2[i] - 1; + //SeedScanState.Ix2[N] = SeedScanState.Ix2[N-1] + 2; + //} + // If 'o', remove midmost and shift quantum numbers by half-integer towards removed one: + + if (whichDSF == 'o') { + for (int i = 0; i < N-1; ++i) + SeedScanState.Ix2[i] = AveragingState.Ix2[i + (i >= N/2)] + 1 - 2*(i >= N/2); + } + // If 'g', add a quantum number in middle (explicitly: to right of index N/2) + // and shift quantum numbers by half-integer away from added one: + if (whichDSF == 'g') { + SeedScanState.Ix2[N/2] = AveragingState.Ix2[N/2] - 1; + for (int i = 0; i < N+1; ++i) + SeedScanState.Ix2[i + (i >= N/2)] = AveragingState.Ix2[i] - 1 + 2*(i >= N/2); + } + SeedScanState.Compute_All(true); + + SeedScanState.Set_Label_from_Ix2 (SeedScanState.Ix2); + + //cout << "which DSF = " << whichDSF << endl; + //cout << "AveragingState Ix2: " << endl << AveragingState.Ix2 << endl; + //cout << "SeedScanState Ix2: " << endl << SeedScanState.Ix2 << endl; + + DP kBT = 0.0; + + // Perform the scan: + General_Scan (whichDSF, iKmin, iKmax, 100000000, kBT, AveragingState, SeedScanState, defaultScanStatename, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + return; + + } + + // Simplified function call of the above: + void Scan_LiebLin (char whichDSF, LiebLin_Bethe_State AveragingState, string defaultScanStatename, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine) + { + int paralevel = 0; + Vect rank(0,1); + Vect nr_processors(0,1); + + Scan_LiebLin (whichDSF, AveragingState, defaultScanStatename, iKmin, iKmax, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + return; + } + + // Scanning on a previously-defined AveragingState + void Scan_Heis (char whichDSF, XXZ_Bethe_State& AveragingState, string defaultScanStatename, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors) + { + // General state scanning for Heisenberg chains + + // PRECONDITIONS: + // - the Ix2 of AveragingState are properly set. + + // Prepare the AveragingState: + AveragingState.Compute_All(true); + + XXZ_Bethe_State SeedScanState; + if (whichDSF == 'Z' || whichDSF == 'z') SeedScanState = AveragingState; + else if (whichDSF == 'm') SeedScanState = Remove_Particle_at_Center (AveragingState); + else if (whichDSF == 'p') SeedScanState = Add_Particle_at_Center (AveragingState); + else JSCerror("Unknown whichDSF in Scan_Heis."); + + //cout << "In General_Scan: SeedScanState = " << SeedScanState << endl; + + // Now the scan itself + General_Scan (whichDSF, iKmin, iKmax, AveragingState.chain.Nsites, 0.0, AveragingState, SeedScanState, defaultScanStatename, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + } + + // Scanning on a previously-defined AveragingState + void Scan_Heis (char whichDSF, XXX_Bethe_State& AveragingState, string defaultScanStatename, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors) + { + // General state scanning for Heisenberg chains + + // PRECONDITIONS: + // - the Ix2 of AveragingState are properly set. + + // Prepare the AveragingState: + AveragingState.Compute_All(true); + + XXX_Bethe_State SeedScanState; + if (whichDSF == 'Z' || whichDSF == 'z') SeedScanState = AveragingState; + else if (whichDSF == 'm') SeedScanState = Remove_Particle_at_Center (AveragingState); + else if (whichDSF == 'p') SeedScanState = Add_Particle_at_Center (AveragingState); + else JSCerror("Unknown whichDSF in Scan_Heis."); + + // Now the scan itself + General_Scan (whichDSF, iKmin, iKmax, AveragingState.chain.Nsites, 0.0, AveragingState, SeedScanState, defaultScanStatename, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + } + + // Scanning on a previously-defined AveragingState + void Scan_Heis (char whichDSF, XXZ_gpd_Bethe_State& AveragingState, string defaultScanStatename, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors) + { + // General state scanning for Heisenberg chains + + // PRECONDITIONS: + // - the Ix2 of AveragingState are properly set. + + // Prepare the AveragingState: + AveragingState.Compute_All(true); + + XXZ_gpd_Bethe_State SeedScanState; + if (whichDSF == 'Z' || whichDSF == 'z') SeedScanState = AveragingState; + else if (whichDSF == 'm') SeedScanState = Remove_Particle_at_Center (AveragingState); + else if (whichDSF == 'p') SeedScanState = Add_Particle_at_Center (AveragingState); + else JSCerror("Unknown whichDSF in Scan_Heis."); + + // Now the scan itself + General_Scan (whichDSF, iKmin, iKmax, AveragingState.chain.Nsites, 0.0, AveragingState, SeedScanState, defaultScanStatename, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + } + + + //void Scan_Heis (char whichDSF, DP Delta, DP N, int M, bool fixed_iK, int iKneeded, + void Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int paralevel, Vect rank, Vect nr_processors) + { + // This function scans the Hilbert space of the Heisenberg spin-1/2 chain + // for the function identified by whichDSF. + + // whichDSF == 'Z': canonical partition function + // whichDSF == 'm': S^{-+} + // whichDSF == 'z': S^{zz} + // whichDSF == 'p': S^{+-} + // whichDSF == 'a': < S^z_j S^z_{j+1} S^z_l S^z_{l+1} > for RIXS + // whichDSF == 'b': < S^z_j S^-_{j+1} S^-_l S^z_{l+1} > + (m <-> z) for RIXS + // whichDSF == 'c': < S^-_j S^-_{j+1} S^-_l S^-_{l+1} > for RIXS + + Heis_Chain BD1(1.0, Delta, 0.0, N); + + Vect_INT Nrapidities_groundstate(0, BD1.Nstrings); + + Nrapidities_groundstate[0] = M; + + Heis_Base baseconfig_groundstate(BD1, Nrapidities_groundstate); + + + if ((Delta > 0.0) && (Delta < 1.0)) { + + XXZ_Bethe_State GroundState(BD1, baseconfig_groundstate); + GroundState.Compute_All(true); + // The ground state is now fully defined. + + XXZ_Bethe_State SeedScanState; + if (whichDSF == 'Z' || whichDSF == 'z') SeedScanState = GroundState; + else if (whichDSF == 'm') SeedScanState = XXZ_Bethe_State(GroundState.chain, M - 1); + else if (whichDSF == 'p') SeedScanState = XXZ_Bethe_State(GroundState.chain, M + 1); + else JSCerror("Unknown whichDSF in Scan_Heis."); + + // Now the scan itself + General_Scan (whichDSF, iKmin, iKmax, N, 0.0, GroundState, SeedScanState, "", Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + } + + else if (Delta == 1.0) { + + XXX_Bethe_State GroundState(BD1, baseconfig_groundstate); + GroundState.Compute_All(true); + // The ground state is now fully defined. + + XXX_Bethe_State SeedScanState; + if (whichDSF == 'Z' || whichDSF == 'z' || whichDSF == 'a' || whichDSF == 'q') SeedScanState = GroundState; + else if (whichDSF == 'm') SeedScanState = XXX_Bethe_State(GroundState.chain, M - 1); + else if (whichDSF == 'p') SeedScanState = XXX_Bethe_State(GroundState.chain, M + 1); + else if (whichDSF == 'c') SeedScanState = XXX_Bethe_State(GroundState.chain, M - 2); + else JSCerror("Unknown whichDSF in Scan_Heis."); + + // Now the scan itself + General_Scan (whichDSF, iKmin, iKmax, N, 0.0, GroundState, SeedScanState, "", Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + } + + else if (Delta > 1.0) { + + XXZ_gpd_Bethe_State GroundState(BD1, baseconfig_groundstate); + GroundState.Compute_All(true); + // The ground state is now fully defined. + + XXZ_gpd_Bethe_State SeedScanState; + if (whichDSF == 'Z' || whichDSF == 'z') SeedScanState = GroundState; + else if (whichDSF == 'm') SeedScanState = XXZ_gpd_Bethe_State(GroundState.chain, M - 1); + else if (whichDSF == 'p') SeedScanState = XXZ_gpd_Bethe_State(GroundState.chain, M + 1); + else JSCerror("Unknown whichDSF in Scan_Heis."); + + // Now the scan itself + General_Scan (whichDSF, iKmin, iKmax, N, 0.0, GroundState, SeedScanState, "", Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + } + + else JSCerror("Delta out of range in Heis_Structure_Factor"); + + return; + } + + void Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine) + { + int paralevel = 0; + Vect rank(0,1); + Vect nr_processors(0,1); + + Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, Max_Secs, target_sumrule, refine, paralevel, rank, nr_processors); + + return; + } + + /* + // Simplified calls: + + void Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, int Max_Secs, bool refine) + { + //Scan_Heis (whichDSF, Delta, N, M, false, 0, Max_Secs, refine, 0, 1); + Scan_Heis (whichDSF, Delta, N, M, iKmin, iKmax, Max_Secs, 1.0e+6, refine, 0, 1); + } + + void Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKneeded, int Max_Secs, bool refine) + { + //Scan_Heis (whichDSF, Delta, N, M, true, iKneeded, Max_Secs, refine, 0, 1); + Scan_Heis (whichDSF, Delta, N, M, iKneeded, iKneeded, Max_Secs, 1.0e+6, refine, 0, 1); + } + + void Scan_Heis (char whichDSF, DP Delta, int N, int M, int Max_Secs, bool refine) + { + //Scan_Heis (whichDSF, Delta, N, M, false, 0, Max_Secs, refine, 0, 1); + Scan_Heis (whichDSF, Delta, N, M, 0, N, Max_Secs, 1.0e+6, refine, 0, 1); + } + */ + /* + void Scan_ODSLF (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + int Max_Secs, DP target_sumrule, bool refine, int rank, int nr_processors) + { + // This function scans the Hilbert space of the spinless fermions related to Heisenberg spin-1/2 chain + // for the function identified by whichDSF. + + // whichDSF == 'Z': canonical partition function + // whichDSF == 'm': S^{-+} + // whichDSF == 'z': S^{zz} + // whichDSF == 'p': S^{+-} + // whichDSF == 'a': < S^z_j S^_{j+1} S^z_l S^z_{l+1} > for RIXS + + Heis_Chain BD1(1.0, Delta, 0.0, N); + + Vect_INT Nrapidities_groundstate(0, BD1.Nstrings); + + Nrapidities_groundstate[0] = M; + + ODSLF_Base baseconfig_groundstate(BD1, Nrapidities_groundstate); + + ODSLF_Ix2_Offsets baseoffsets(baseconfig_groundstate, 0ULL); + + if ((Delta > 0.0) && (Delta < 1.0)) { + + ODSLF_XXZ_Bethe_State GroundState(BD1, baseconfig_groundstate); + GroundState.Compute_All(true); + + // The ground state is now fully defined. Now the scan itself + //General_Scan (whichDSF, fixed_iK, iKneeded, N, GroundState, GroundState, Max_Secs, refine, rank, nr_processors); + General_Scan (whichDSF, iKmin, iKmax, N, 0.0, GroundState, GroundState, Max_Secs, target_sumrule, refine, rank, nr_processors); + + } + */ + /* + else if (Delta == 1.0) { + + XXX_Bethe_State GroundState(BD1, baseconfig_groundstate); + GroundState.Compute_All(true); + + // The ground state is now fully defined. Now the scan itself + //General_Scan (whichDSF, fixed_iK, iKneeded, N, GroundState, GroundState, Max_Secs, refine, rank, nr_processors); + General_Scan (whichDSF, iKmin, iKmax, N, GroundState, GroundState, Max_Secs, refine, rank, nr_processors); + } + + else if (Delta > 1.0) { + + XXZ_gpd_Bethe_State GroundState(BD1, baseconfig_groundstate); + GroundState.Compute_All(true); + + // The ground state is now fully defined. Now the scan itself + //General_Scan (whichDSF, fixed_iK, iKneeded, N, GroundState, GroundState, Max_Secs, refine, rank, nr_processors); + General_Scan (whichDSF, iKmin, iKmax, N, GroundState, GroundState, Max_Secs, refine, rank, nr_processors); + } + */ + /* + else JSCerror("Delta out of range in ODSLF Structure Factor"); + + return; + } + + // Simplified calls: + + void Scan_ODSLF (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, int Max_Secs, bool refine) + { + //Scan_Heis (whichDSF, Delta, N, M, false, 0, Max_Secs, refine, 0, 1); + Scan_ODSLF (whichDSF, Delta, N, M, iKmin, iKmax, Max_Secs, 1.0e+6, refine, 0, 1); + } + + void Scan_ODSLF (char whichDSF, DP Delta, int N, int M, int iKneeded, int Max_Secs, bool refine) + { + //Scan_Heis (whichDSF, Delta, N, M, true, iKneeded, Max_Secs, refine, 0, 1); + Scan_ODSLF (whichDSF, Delta, N, M, iKneeded, iKneeded, Max_Secs, 1.0e+6, refine, 0, 1); + } + + void Scan_ODSLF (char whichDSF, DP Delta, int N, int M, int Max_Secs, bool refine) + { + //Scan_Heis (whichDSF, Delta, N, M, false, 0, Max_Secs, refine, 0, 1); + Scan_ODSLF (whichDSF, Delta, N, M, 0, N, Max_Secs, 1.0e+6, refine, 0, 1); + } + + + + // Geometric quenches + + void Scan_LiebLin_Geometric_Quench (DP c_int, DP L_1, int type_id_1, long long int id_1, DP L_2, int N, int iK_UL, + int Max_Secs, DP target_sumrule, bool refine) + { + // We decompose the wavefunction of state 1 (living on length L_1) into + // the wavefunctions living on length L_2. + + // IMPORTANT ASSUMPTIONS: + + LiebLin_Bethe_State lstate(c_int, L_1, N, iK_UL, type_id_1); + lstate.Set_to_id(id_1); + lstate.Compute_All(true); + + // We now put the rapidities and norm into a state in length L_2, + // which will serve as basis for the scan. + + LiebLin_Bethe_State lstate2(c_int, L_2, N, iK_UL, type_id_1); + lstate2.Set_to_id (0LL); + lstate2.Compute_All(true); + + char whichDSF = 'q'; + + //General_Scan (whichDSF, false, 0, 100000000, lstate, lstate2, Max_Secs, refine, 0, 1); + General_Scan (whichDSF, -iK_UL, iK_UL, 100000000, 0.0, lstate, lstate2, Max_Secs, target_sumrule, refine, 0, 1); + + return; + } + + + void Scan_Heis_Geometric_Quench (DP Delta, int N_1, int M, long long int base_id_1, long long int type_id_1, long long int id_1, + int N_2, int iKmin, int iKmax, int Max_Secs, DP target_sumrule, bool refine) + { + // We decompose the wavefunction of state 1 (living on length L_1) into + // the wavefunctions living on length L_2. + + Heis_Chain BD_1(1.0, Delta, 0.0, N_1); + Heis_Chain BD_2(1.0, Delta, 0.0, N_2); + + + if ((Delta > 0.0) && (Delta < 1.0)) { + JSCerror("Geometric quench not yet implemented for XXZ."); + } + + else if (Delta == 1.0) { + + XXX_Bethe_State BasicState_1(BD_1, base_id_1, type_id_1); + BasicState_1.Set_to_id (id_1); + BasicState_1.Compute_All(true); + + // Ref state for scanning: + XXX_Bethe_State BasicState_2(BD_2, M); + BasicState_2.Set_to_id (0LL); + BasicState_2.Compute_All(true); + + char whichDSF = 'q'; + + // The ground state is now fully defined. Now the scan itself + //General_Scan (whichDSF, fixed_iK, iKneeded, N, GroundState, GroundState, Max_Secs, refine, rank, nr_processors); + General_Scan (whichDSF, iKmin, iKmax, N_2, 0.0, BasicState_1, BasicState_2, Max_Secs, target_sumrule, refine, 0, 1); + } + + else if (Delta > 1.0) { + JSCerror("Geometric quench not yet implemented for XXZ_gpd."); + } + + else JSCerror("Delta out of range in Heis_Structure_Factor"); + + return; + } + */ + +} // namespace JSC diff --git a/src/SCAN/General_Scan_Parallel.cc b/src/SCAN/General_Scan_Parallel.cc new file mode 100644 index 0000000..bcfea62 --- /dev/null +++ b/src/SCAN/General_Scan_Parallel.cc @@ -0,0 +1,782 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/SCAN/General_Scan_Parallel.cc + +Purpose: universal implementation of state scanning: + functions to descend down hierarchy of intermediate states. + + Parallel implementation using MPI. + +NOTE: since templated functions have to be in the same file, + we put all scanning functions here. The externally-used + functions are defined at the end of this file. + + +***********************************************************/ + +//#include "mpi.h" +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + //*******************************************************************// + // Functions applicable to all calculations: + + void Split_thr_Files (string prefix, char whichDSF, int nr_processors_at_newlevel) + { + // From an existing threads file directory, this produces nr_processors_at_newlevel thread file directories + // containing distributed threads. + + stringstream THRDIR_stringstream; string THRDIR_string; + THRDIR_stringstream << prefix << "_thrdir"; + THRDIR_string = THRDIR_stringstream.str(); + + // Load all the info about the threads, in particular the nthreads_total vector: + Scan_Thread_Data thr_data(THRDIR_string, true); + thr_data.Load(); + + // Create the different directories: + Vect THRDIRS_stringstream(nr_processors_at_newlevel); + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + THRDIRS_stringstream[rank] << prefix << "_" << rank << "_" << nr_processors_at_newlevel << "_thrdir"; + mkdir((THRDIRS_stringstream[rank].str()).c_str(), S_IRWXU | S_IRWXG | S_IRWXO); + } + + Vect thr_data_par(nr_processors_at_newlevel); + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) + thr_data_par[rank] = Scan_Thread_Data (THRDIRS_stringstream[rank].str(), false); // put refine == false here to avoid loading any deprecated data + + // Transfer all the existing threads into the new ones: + int rankindex = 0; + int il; + while ((il = thr_data.lowest_il_with_nthreads_neq_0) < thr_data.nlists) { + Vect next_threads = thr_data.Extract_Next_Scan_Threads(); + for (int it = 0; it < next_threads.size(); ++it) { + thr_data_par[rankindex++].Include_Thread (il, next_threads[it].label, next_threads[it].type); + rankindex = rankindex % nr_processors_at_newlevel; + } + if (il == thr_data.nlists - 1) break; + // Flush data to disk to preserve memory + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) + thr_data_par[rank].Flush_to_Disk(il); + } + // thr_data is now empty + thr_data.Save(); + + // remove the nthreads.dat file from original threads directory: + stringstream datfile_strstream; + datfile_strstream << THRDIR_string << "/nthreads.dat"; + string datfilename = datfile_strstream.str(); + remove(datfilename.c_str()); + + // Save the threads in the new directories: + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) thr_data_par[rank].Save(); + + return; + } + + /* + void Create_Empty_Files (string prefix, char whichDSF, int nr_processors_at_newlevel) + { + // This function creates, for convenience, a set of 'empty' files, so a full set of files is available at all paralevels. + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream RAW_stringstream; string RAW_string; + stringstream INADM_stringstream; string INADM_string; + stringstream CONV0_stringstream; string CONV0_string; + stringstream LOG_stringstream; string LOG_string; + //stringstream THR_stringstream; string THR_string; + stringstream SRC_stringstream; string SRC_string; + stringstream FSR_stringstream; string FSR_string; + stringstream SUM_stringstream; string SUM_string; + + RAW_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".raw"; + INADM_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".inadm"; + CONV0_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".conv0"; + LOG_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".log"; + //THR_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".thr"; + SRC_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".src"; + FSR_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".fsr"; + SUM_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".sum"; + + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + INADM_string = INADM_stringstream.str(); const char* INADM_Cstr = INADM_string.c_str(); + CONV0_string = CONV0_stringstream.str(); const char* CONV0_Cstr = CONV0_string.c_str(); + LOG_string = LOG_stringstream.str(); const char* LOG_Cstr = LOG_string.c_str(); + //THR_string = THR_stringstream.str(); const char* THR_Cstr = THR_string.c_str(); + SRC_string = SRC_stringstream.str(); const char* SRC_Cstr = SRC_string.c_str(); + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + SUM_string = SUM_stringstream.str(); const char* SUM_Cstr = SUM_string.c_str(); + + // We open and close these files (except for SUM, which we fill with a zero-valued scan_info + fstream RAW_file; RAW_file.open(RAW_Cstr); RAW_file.close(); + fstream INADM_file; INADM_file.open(INADM_Cstr); INADM_file.close(); + fstream CONV0_file; CONV0_file.open(CONV0_Cstr); CONV0_file.close(); + fstream LOG_file; LOG_file.open(LOG_Cstr); LOG_file.close(); + Scan_Info emptyinfo; emptyinfo.Save(SRC_Cstr); + fstream FSR_file; FSR_file.open(FSR_Cstr); FSR_file.close(); + fstream SUM_file; SUM_file.open(SUM_Cstr); SUM_file.close(); + + } + } + */ + + void Merge_raw_Files (string prefix, char whichDSF, int nr_processors_at_newlevel) + { + + // Open the original raw file: + stringstream RAW_stringstream; string RAW_string; + //RAW_stringstream << prefix; + //if (whichDSF == 'Z') RAW_stringstream << ".dat"; + //else RAW_stringstream << ".raw"; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + fstream RAW_outfile; + RAW_outfile.open(RAW_Cstr, fstream::out | fstream::app); // NB: we append !! + if (RAW_outfile.fail()) JSCerror("Could not open RAW_outfile... "); + RAW_outfile.precision(16); + + // Append all other raw files to original one + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream RAW_in_stringstream; string RAW_in_string; + RAW_in_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".raw"; + RAW_in_string = RAW_in_stringstream.str(); const char* RAW_in_Cstr = RAW_in_string.c_str(); + + ifstream RAW_infile; + RAW_infile.open(RAW_in_Cstr); + if (RAW_infile.fail()) { + //cout << RAW_in_Cstr << endl; + //JSCerror ("Could not open file."); + continue; // if file isn't there, just continue... + } + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + int nr, nl; + while (RAW_infile.peek() != EOF) { + //RAW_infile >> omega >> iK >> FF >> conv >> label; + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (whichDSF == '1') RAW_infile >> nr >> nl; + //RAW_outfile << endl << omega << "\t" << iK << "\t" << FF << "\t" << conv << "\t" << label; + RAW_outfile << endl << omega << "\t" << iK << "\t" << FF << "\t" << dev << "\t" << label; + if (whichDSF == '1') RAW_outfile << "\t" << nr << "\t" << nl; + } + + RAW_infile.close(); + + // Delete file + remove(RAW_in_Cstr); + + } // for rank + + RAW_outfile.close(); + + return; + } + + void Merge_thr_Files (string prefix, char whichDSF, int nr_processors_at_newlevel) + { + + stringstream THRDIR_stringstream; string THRDIR_string; + THRDIR_stringstream << prefix << "_thrdir"; + THRDIR_string = THRDIR_stringstream.str(); + + Scan_Thread_Data thr_data(THRDIR_string, true); + + Vect THRDIRS_stringstream(nr_processors_at_newlevel); + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) + THRDIRS_stringstream[rank] << prefix << "_" << rank << "_" << nr_processors_at_newlevel << "_thrdir"; + + Vect thr_data_par(nr_processors_at_newlevel); + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + thr_data_par[rank] = Scan_Thread_Data (THRDIRS_stringstream[rank].str(), true); + thr_data_par[rank].Load(); + } + + // Transfer from all the existing threads into the base one: + int il; + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + while ((il = thr_data_par[rank].lowest_il_with_nthreads_neq_0) < thr_data_par[rank].nlists) { + Vect next_threads = thr_data_par[rank].Extract_Next_Scan_Threads(); + for (int it = 0; it < next_threads.size(); ++it) { + thr_data.Include_Thread (il, next_threads[it].label, next_threads[it].type); + } + if (il == thr_data_par[rank].nlists - 1) break; + } + } + // all thr_data_par are now empty + // remove the nthreads.dat files from original threads directory: + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream datfile_strstream; + datfile_strstream << THRDIRS_stringstream[rank].str() << "/nthreads.dat"; + string datfilename = datfile_strstream.str(); + remove(datfilename.c_str()); + rmdir((THRDIRS_stringstream[rank].str()).c_str()); + } + + thr_data.Save(); + + return; + } + + template + void Merge_sum_Files (Tstate& SeedScanState, string prefix, char whichDSF, int nr_processors_at_newlevel) + { + + Scan_State_List ScanStateList (whichDSF, SeedScanState); + ScanStateList.Populate_List(whichDSF, SeedScanState); + + // Open the original file: + stringstream SUM_stringstream; string SUM_string; + SUM_stringstream << prefix << ".sum"; + SUM_string = SUM_stringstream.str(); const char* SUM_Cstr = SUM_string.c_str(); + + // Load the original info: + //ScanStateList.Load_Info (SUM_Cstr); // Not needed anymore: rank 0 has loaded the original info + if (file_exists(SUM_Cstr)) ScanStateList.Load_Info (SUM_Cstr); // Needed again! + + // Load all other info: + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream SUM_in_stringstream; string SUM_in_string; + SUM_in_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".sum"; + SUM_in_string = SUM_in_stringstream.str(); const char* SUM_in_Cstr = SUM_in_string.c_str(); + + if (!file_exists(SUM_in_Cstr)) continue; + + ScanStateList.Load_Info (SUM_in_Cstr); + + // Delete file + remove(SUM_in_Cstr); + + } // for rank + + // Save the result in original sum file: + ScanStateList.Order_in_SRC (); + ScanStateList.Save_Info(SUM_Cstr); + + return; + } + + void Merge_inadm_conv0_src_stat_log_Files (string prefix, char whichDSF, int nr_processors_at_newlevel) + { + // This also deletes any .fsr files at the newlevel. + + // Open the original src file: + stringstream SRC_stringstream; string SRC_string; + SRC_stringstream << prefix << ".src"; + SRC_string = SRC_stringstream.str(); const char* SRC_Cstr = SRC_string.c_str(); + + // Load original info: + Scan_Info scan_info; + // If file is there, load it: + if (file_exists(SRC_Cstr)) scan_info.Load(SRC_Cstr); + Scan_Info scan_info_before = scan_info; + + // Load all other info: + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream SRC_in_stringstream; string SRC_in_string; + SRC_in_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".src"; + SRC_in_string = SRC_in_stringstream.str(); const char* SRC_in_Cstr = SRC_in_string.c_str(); + + if (!file_exists(SRC_in_Cstr)) continue; + + Scan_Info scan_info_in; + scan_info_in.Load (SRC_in_Cstr); + + scan_info += scan_info_in; + + // Delete file + remove(SRC_in_Cstr); + + } // for rank + + // Save the result in original src file: + scan_info.Save(SRC_Cstr); + + // Load content of all other inadm files into main inadm: + stringstream INADM_stringstream; string INADM_string; + INADM_stringstream << prefix << ".inadm"; + INADM_string = INADM_stringstream.str(); const char* INADM_Cstr = INADM_string.c_str(); + ofstream INADM_outfile; + INADM_outfile.open(INADM_Cstr, fstream::out | fstream::app); + if (INADM_outfile.fail()) JSCerror("Could not open INADM_outfile... "); + INADM_outfile.precision(16); + + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream INADM_in_stringstream; string INADM_in_string; + INADM_in_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".inadm"; + INADM_in_string = INADM_in_stringstream.str(); const char* INADM_in_Cstr = INADM_in_string.c_str(); + + if (!file_exists(INADM_in_Cstr)) continue; + + ifstream INADM_infile; + INADM_infile.open(INADM_in_Cstr); + + string line; + string linerej; + while (INADM_infile.good() ) { + getline (INADM_infile, line); + if (line.compare("") != 0) INADM_outfile << line << endl; + } + + INADM_infile.close(); + + // Delete file + remove(INADM_in_Cstr); + + } // for rank + + INADM_outfile.close(); + + // Load content of all other conv0 files into main inadm: + stringstream CONV0_stringstream; string CONV0_string; + CONV0_stringstream << prefix << ".conv0"; + CONV0_string = CONV0_stringstream.str(); const char* CONV0_Cstr = CONV0_string.c_str(); + ofstream CONV0_outfile; + CONV0_outfile.open(CONV0_Cstr, fstream::out | fstream::app); + if (CONV0_outfile.fail()) JSCerror("Could not open CONV0_outfile... "); + CONV0_outfile.precision(16); + + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream CONV0_in_stringstream; string CONV0_in_string; + CONV0_in_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".conv0"; + CONV0_in_string = CONV0_in_stringstream.str(); const char* CONV0_in_Cstr = CONV0_in_string.c_str(); + + if (!file_exists(CONV0_in_Cstr)) continue; + + ifstream CONV0_infile; + CONV0_infile.open(CONV0_in_Cstr); + + string line; + string linerej; + while (CONV0_infile.good() ) { + getline (CONV0_infile, line); + if (line.compare("") != 0) CONV0_outfile << line << endl; + } + + CONV0_infile.close(); + + // Delete file + remove(CONV0_in_Cstr); + + } // for rank + + CONV0_outfile.close(); + + // Load content of all other stat files into main: + stringstream STAT_stringstream; string STAT_string; + STAT_stringstream << prefix << ".stat"; + STAT_string = STAT_stringstream.str(); const char* STAT_Cstr = STAT_string.c_str(); + ofstream STAT_outfile; + STAT_outfile.open(STAT_Cstr, fstream::out | fstream::app); + if (STAT_outfile.fail()) JSCerror("Could not open STAT_outfile... "); + STAT_outfile.precision(16); + + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream STAT_in_stringstream; string STAT_in_string; + STAT_in_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".stat"; + STAT_in_string = STAT_in_stringstream.str(); const char* STAT_in_Cstr = STAT_in_string.c_str(); + + if (!file_exists(STAT_in_Cstr)) continue; + + ifstream STAT_infile; + STAT_infile.open(STAT_in_Cstr); + + string line; + string linerej; + while (STAT_infile.good() ) { + getline (STAT_infile, line); + if (line.compare("") != 0) STAT_outfile << line << endl; + } + + STAT_infile.close(); + + // Delete file + remove(STAT_in_Cstr); + + } // for rank + + STAT_outfile.close(); + + + // Put some digested info in log file: + Scan_Info scan_info_refinement; + scan_info_refinement = scan_info; + scan_info_refinement -= scan_info_before; + stringstream LOG_stringstream; string LOG_string; + LOG_stringstream << prefix << ".log"; + LOG_string = LOG_stringstream.str(); const char* LOG_Cstr = LOG_string.c_str(); + ofstream LOG_outfile; + LOG_outfile.open(LOG_Cstr, fstream::out | fstream::app); + if (LOG_outfile.fail()) JSCerror("Could not open LOG_outfile... "); + LOG_outfile.precision(16); + + // Load content of all other log files into main log: + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream LOG_in_stringstream; string LOG_in_string; + LOG_in_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".log"; + LOG_in_string = LOG_in_stringstream.str(); const char* LOG_in_Cstr = LOG_in_string.c_str(); + + if (!file_exists(LOG_in_Cstr)) continue; + + ifstream LOG_infile; + LOG_infile.open(LOG_in_Cstr); + + string line; + while (LOG_infile.good() ) { + getline (LOG_infile, line); + LOG_outfile << line << endl; + } + + LOG_infile.close(); + + // Delete file + remove(LOG_in_Cstr); + + } // for rank + + // Now remove the .fsr files at the newlevel: + for (int rank = 0; rank < nr_processors_at_newlevel; ++rank) { + stringstream FSR_in_stringstream; string FSR_in_string; + FSR_in_stringstream << prefix << "_" << rank << "_" << nr_processors_at_newlevel << ".fsr"; + FSR_in_string = FSR_in_stringstream.str(); const char* FSR_in_Cstr = FSR_in_string.c_str(); + + // Delete file + remove(FSR_in_Cstr); + } + + LOG_outfile << "Refining in parallel mode using " << nr_processors_at_newlevel << " processors." + << endl << "Refining info: " << scan_info_refinement + << endl << "Resulting info: " << scan_info << endl; + LOG_outfile << "Code version " << JSC_VERSION << ", copyright J.-S. Caux." << endl; + LOG_outfile.close(); + + return; + } + + + //****************************************************************************// + // Model-specific functions: + + //void Prepare_Parallel_Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iK_UL, bool fixed_iK, int iKneeded, + void Prepare_Parallel_Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, + string defaultScanStatename, + //int Max_Secs, bool refine, int rank, + int paralevel, Vect rank_lower_paralevels, Vect nr_processors_lower_paralevels, + int nr_processors_at_newlevel) + { + // From an existing scan, this function splits the threads into + // nr_processors_at_newlevel separate files, from which the parallel process + // can be launched. + + LiebLin_Bethe_State GroundState (c_int, L, N); + + // Define file name + stringstream filenameprefix; + //Data_File_Name (filenameprefix, whichDSF, fixed_iK, iKneeded, GroundState, GroundState); + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, kBT, GroundState, GroundState, defaultScanStatename); + for (int i = 0; i < paralevel - 1; ++i) filenameprefix << "_" << rank_lower_paralevels[i] << "_" << nr_processors_lower_paralevels[i]; + string prefix = filenameprefix.str(); + + Split_thr_Files (prefix, whichDSF, nr_processors_at_newlevel); + + //Create_Empty_Files (prefix, whichDSF, nr_processors_at_newlevel); + + return; + } + + //void Wrapup_Parallel_Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iK_UL, bool fixed_iK, int iKneeded, + void Wrapup_Parallel_Scan_LiebLin (char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, + string defaultScanStatename, + //int Max_Secs, bool refine, int rank, + int paralevel, Vect rank_lower_paralevels, Vect nr_processors_lower_paralevels, + int nr_processors_at_newlevel) + { + //DP epsilon = log(L)/L; + + //LiebLin_Bethe_State GroundState (c_int, L, N); + //LiebLin_Bethe_State spstate = Canonical_Saddle_Point_State (c_int, L, N, kBT, epsilon); + //LiebLin_Bethe_State spstate = Canonical_Saddle_Point_State (c_int, L, N, kBT); + + // Read the saddle-point state from the sps file: + stringstream SPS_stringstream; string SPS_string; + //SPS_stringstream << "Tgt0_"; + //Data_File_Name (SPS_stringstream, whichDSF, iKmin, iKmax, kBT, spstate, SeedScanState, ""); + Data_File_Name (SPS_stringstream, whichDSF, c_int, L, N, iKmin, iKmax, kBT, 0.0, ""); + SPS_stringstream << ".sps"; + SPS_string = SPS_stringstream.str(); const char* SPS_Cstr = SPS_string.c_str(); + + LiebLin_Bethe_State spstate; + + if (file_exists(SPS_Cstr)) { + + fstream spsfile; + spsfile.open(SPS_Cstr, fstream::in); + + int Nspsread; + spsfile >> Nspsread; + if (Nspsread != N) { + cout << Nspsread << "\t" << N << endl; + JSCerror("Wrong number of Ix2 in saddle-point state."); + } + spstate = LiebLin_Bethe_State (c_int, L, N); + for (int i = 0; i < N; ++i) spsfile >> spstate.Ix2[i]; + + spsfile.close(); + + spstate.Compute_All(true); + } + + else spstate = Canonical_Saddle_Point_State (c_int, L, N, kBT); + + int Nscan = N; + if (whichDSF == 'o') Nscan = N - 1; + if (whichDSF == 'g') Nscan = N + 1; + + LiebLin_Bethe_State SeedScanState = spstate; + //if (whichDSF == 'o' || whichDSF == 'g') SeedScanState = Canonical_Saddle_Point_State (c_int, L, Nscan, kBT, epsilon); + if (whichDSF == 'o' || whichDSF == 'g') SeedScanState = Canonical_Saddle_Point_State (c_int, L, Nscan, kBT); + + // Define file name + stringstream filenameprefix; + //Data_File_Name (filenameprefix, whichDSF, fixed_iK, iKneeded, GroundState, GroundState); + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, kBT, spstate, SeedScanState, defaultScanStatename); + for (int i = 0; i < paralevel - 1; ++i) filenameprefix << "_" << rank_lower_paralevels[i] << "_" << nr_processors_lower_paralevels[i]; + string prefix = filenameprefix.str(); + + + // Merge raw files + Merge_raw_Files (prefix, whichDSF, nr_processors_at_newlevel); + + // Merge thr files + Merge_thr_Files (prefix, whichDSF, nr_processors_at_newlevel); + + // Merge sum files + Merge_sum_Files (SeedScanState, prefix, whichDSF, nr_processors_at_newlevel); + + // Merge src files + Merge_inadm_conv0_src_stat_log_Files (prefix, whichDSF, nr_processors_at_newlevel); + // This also puts some digested info in log file. + + // Evaluate f-sumrule: + /* + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + stringstream FSR_stringstream; string FSR_string; + FSR_stringstream << prefix << ".fsr"; + FSR_string = FSR_stringstream.str(); const char* FSR_Cstr = FSR_string.c_str(); + */ + DP Chem_Pot = Chemical_Potential (spstate); + //if (!fixed_iK) if (whichDSF != 'q') Evaluate_F_Sumrule (whichDSF, GroundState, Chem_Pot, RAW_Cstr, FSR_Cstr); + //if (iKmin != iKmax) if (whichDSF != 'q') Evaluate_F_Sumrule (whichDSF, GroundState, Chem_Pot, iKmin, iKmax, RAW_Cstr, FSR_Cstr); + if (iKmin != iKmax) if (whichDSF != 'q') Evaluate_F_Sumrule (prefix, whichDSF, spstate, Chem_Pot, iKmin, iKmax); + + // ... and we're done. + + return; + } + + + //****************************************************************************// + // Heisenberg: + + void Prepare_Parallel_Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + //int Max_Secs, bool refine, int rank, + int paralevel, Vect rank_lower_paralevels, Vect nr_processors_lower_paralevels, + int nr_processors_at_newlevel) + { + // From an existing scan, this function splits the threads into + // nr_processors separate files, from which the parallel process + // can be launched. + + Heis_Chain BD1(1.0, Delta, 0.0, N); + + Vect_INT Nrapidities_groundstate(0, BD1.Nstrings); + + Nrapidities_groundstate[0] = M; + + Heis_Base baseconfig_groundstate(BD1, Nrapidities_groundstate); + + //Ix2_Offsets baseoffsets(baseconfig_groundstate, 0ULL); + + // Define file name + stringstream filenameprefix; + + if ((Delta > 0.0) && (Delta < 1.0)) { + + XXZ_Bethe_State GroundState(BD1, baseconfig_groundstate); + + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, 0.0, GroundState, GroundState, ""); + } + + else if (Delta == 1.0) { + + XXX_Bethe_State GroundState(BD1, baseconfig_groundstate); + + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, 0.0, GroundState, GroundState, ""); + } + + else if (Delta > 1.0) { + + XXZ_gpd_Bethe_State GroundState(BD1, baseconfig_groundstate); + + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, 0.0, GroundState, GroundState, ""); + } + + else JSCerror("Delta out of range in Prepare_Parallel_Scan_Heis"); + + for (int i = 0; i < paralevel - 1; ++i) filenameprefix << "_" << rank_lower_paralevels[i] << "_" << nr_processors_lower_paralevels[i]; + string prefix = filenameprefix.str(); + + Split_thr_Files (prefix, whichDSF, nr_processors_at_newlevel); + + return; + } + + + void Wrapup_Parallel_Scan_Heis (char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, + int paralevel, Vect rank_lower_paralevels, Vect nr_processors_lower_paralevels, + int nr_processors_at_newlevel) + { + // From an existing scan, this function splits the threads into + // nr_processors separate files, from which the parallel process + // can be launched. + + Heis_Chain BD1(1.0, Delta, 0.0, N); + + Vect_INT Nrapidities_groundstate(0, BD1.Nstrings); + + Nrapidities_groundstate[0] = M; + + Heis_Base baseconfig_groundstate(BD1, Nrapidities_groundstate); + + //Ix2_Offsets baseoffsets(baseconfig_groundstate, 0ULL); + + // Define file name + stringstream filenameprefix; + string prefix; + + if ((Delta > 0.0) && (Delta < 1.0)) { + + XXZ_Bethe_State GroundState(BD1, baseconfig_groundstate); + + XXZ_Bethe_State SeedScanState; + if (whichDSF == 'Z' || whichDSF == 'z') SeedScanState = GroundState; + else if (whichDSF == 'm') SeedScanState = XXZ_Bethe_State(GroundState.chain, M - 1); + else if (whichDSF == 'p') SeedScanState = XXZ_Bethe_State(GroundState.chain, M + 1); + else JSCerror("Unknown whichDSF in Scan_Heis."); + + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, 0.0, GroundState, SeedScanState, ""); + for (int i = 0; i < paralevel - 1; ++i) filenameprefix << "_" << rank_lower_paralevels[i] << "_" << nr_processors_lower_paralevels[i]; + prefix = filenameprefix.str(); + + // Merge sum files + Merge_sum_Files (SeedScanState, prefix, whichDSF, nr_processors_at_newlevel); + + // Merge raw files + Merge_raw_Files (prefix, whichDSF, nr_processors_at_newlevel); + + // Merge thr files + Merge_thr_Files (prefix, whichDSF, nr_processors_at_newlevel); + + // Merge src files + Merge_inadm_conv0_src_stat_log_Files (prefix, whichDSF, nr_processors_at_newlevel); + // This also puts some digested info in log file. + + // Evaluate f-sumrule: + DP Chem_Pot = Chemical_Potential (GroundState); + if (iKmin != iKmax) if (whichDSF != 'q') Evaluate_F_Sumrule (prefix, whichDSF, GroundState, Chem_Pot, iKmin, iKmax); + + } + + else if (Delta == 1.0) { + + XXX_Bethe_State GroundState(BD1, baseconfig_groundstate); + + XXX_Bethe_State SeedScanState; + if (whichDSF == 'Z' || whichDSF == 'z' || whichDSF == 'a' || whichDSF == 'q') SeedScanState = GroundState; + else if (whichDSF == 'm') SeedScanState = XXX_Bethe_State(GroundState.chain, M - 1); + else if (whichDSF == 'p') SeedScanState = XXX_Bethe_State(GroundState.chain, M + 1); + else if (whichDSF == 'c') SeedScanState = XXX_Bethe_State(GroundState.chain, M - 2); + else JSCerror("Unknown whichDSF in Scan_Heis."); + + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, 0.0, GroundState, SeedScanState, ""); + for (int i = 0; i < paralevel - 1; ++i) filenameprefix << "_" << rank_lower_paralevels[i] << "_" << nr_processors_lower_paralevels[i]; + prefix = filenameprefix.str(); + + // Merge sum files + Merge_sum_Files (SeedScanState, prefix, whichDSF, nr_processors_at_newlevel); + + // Merge raw files + Merge_raw_Files (prefix, whichDSF, nr_processors_at_newlevel); + + // Merge thr files + Merge_thr_Files (prefix, whichDSF, nr_processors_at_newlevel); + + // Merge src files + Merge_inadm_conv0_src_stat_log_Files (prefix, whichDSF, nr_processors_at_newlevel); + // This also puts some digested info in log file. + + // Evaluate f-sumrule: + DP Chem_Pot = Chemical_Potential (GroundState); + if (iKmin != iKmax) if (whichDSF != 'q') Evaluate_F_Sumrule (prefix, whichDSF, GroundState, Chem_Pot, iKmin, iKmax); + + } + + else if (Delta > 1.0) { + + XXZ_gpd_Bethe_State GroundState(BD1, baseconfig_groundstate); + + XXZ_gpd_Bethe_State SeedScanState; + if (whichDSF == 'Z' || whichDSF == 'z') SeedScanState = GroundState; + else if (whichDSF == 'm') SeedScanState = XXZ_gpd_Bethe_State(GroundState.chain, M - 1); + else if (whichDSF == 'p') SeedScanState = XXZ_gpd_Bethe_State(GroundState.chain, M + 1); + else JSCerror("Unknown whichDSF in Scan_Heis."); + + Data_File_Name (filenameprefix, whichDSF, iKmin, iKmax, 0.0, GroundState, SeedScanState, ""); + for (int i = 0; i < paralevel - 1; ++i) filenameprefix << "_" << rank_lower_paralevels[i] << "_" << nr_processors_lower_paralevels[i]; + prefix = filenameprefix.str(); + + // Merge sum files + Merge_sum_Files (SeedScanState, prefix, whichDSF, nr_processors_at_newlevel); + + // Merge raw files + Merge_raw_Files (prefix, whichDSF, nr_processors_at_newlevel); + + // Merge thr files + Merge_thr_Files (prefix, whichDSF, nr_processors_at_newlevel); + + // Merge src files + Merge_inadm_conv0_src_stat_log_Files (prefix, whichDSF, nr_processors_at_newlevel); + // This also puts some digested info in log file. + + // Evaluate f-sumrule: + DP Chem_Pot = Chemical_Potential (GroundState); + if (iKmin != iKmax) if (whichDSF != 'q') Evaluate_F_Sumrule (prefix, whichDSF, GroundState, Chem_Pot, iKmin, iKmax); + + } + + else JSCerror("Delta out of range in Prepare_Parallel_Scan_Heis"); + + + // ... and we're done. + + return; + } + +} // namespace JSC diff --git a/src/SCAN/Particle_Hole_Excitation_Cost.cc b/src/SCAN/Particle_Hole_Excitation_Cost.cc new file mode 100644 index 0000000..20d7e3f --- /dev/null +++ b/src/SCAN/Particle_Hole_Excitation_Cost.cc @@ -0,0 +1,56 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/SCAN/Particle_Hole_Excitation_Cost.cc + +Purpose: defines the expected cost of adding a p-h pair. Used in General_Scan. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + DP Particle_Hole_Excitation_Cost (char whichDSF, LiebLin_Bethe_State& AveragingState) + { + // Estimates the cost of adding a particle-hole excitation to an intermediate state + DP ph_cost = 1.0; + + //if (whichDSF == 'd') ph_cost = JSC::min(0.1, 1.0/sqrt(AveragingState.c_int)); + //if (whichDSF == 'd') ph_cost = JSC::min(0.1, 1.0/AveragingState.c_int); + if (whichDSF == 'd') ph_cost = JSC::min(0.01, 0.1/AveragingState.c_int); + //if (whichDSF == 'd') ph_cost = JSC::min(0.001, 0.01/AveragingState.c_int); + else if (whichDSF == 'o') ph_cost = 0.01; + else if (whichDSF == 'g') ph_cost = 0.01; + else if (whichDSF == 'Z') ph_cost = 1.0; + else if (whichDSF == 'B') ph_cost = 0.1; + else if (whichDSF == 'C') ph_cost = 0.1; + else JSCerror("whichDSF option not implemented in Particle_Hole_Excitation_Cost for LiebLin."); + + return(ph_cost); + } + + DP Particle_Hole_Excitation_Cost (char whichDSF, Heis_Bethe_State& AveragingState) + { + // Estimates the cost of adding a particle-hole excitation to an intermediate state + DP ph_cost = 1.0; + + if (whichDSF == 'z') ph_cost = 0.01; + else if (whichDSF == 'm') ph_cost = 0.01; + else if (whichDSF == 'p') ph_cost = 0.01; + else if (whichDSF == 'Z') ph_cost = 1.0; + else JSCerror("whichDSF option not implemented in Particle_Hole_Excitation_Cost for HEIS."); + + return(ph_cost); + } + +} // namespace JSC + diff --git a/src/SCAN/Scan_Info.cc b/src/SCAN/Scan_Info.cc new file mode 100644 index 0000000..8d79261 --- /dev/null +++ b/src/SCAN/Scan_Info.cc @@ -0,0 +1,97 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Scan_Info.cc + +Purpose: defines all functions for Scan_Info class. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + Scan_Info::Scan_Info() : + //sumrule_obtained(0.0), Nfull(0LL), Ninadm(0LL), Ndata(0LL), Ndata_conv(0LL), Ndata_conv0(0LL), CPU_ticks(0LL), CPU_ticks_TOT(0LL) {} + //sumrule_obtained(0.0), Nfull(0.0), Ninadm(0LL), Ndata(0LL), Ndata_conv(0LL), Ndata_conv0(0LL), CPU_ticks(0LL), CPU_ticks_TOT(0LL) {} + sumrule_obtained(0.0), Nfull(0.0), Ninadm(0LL), Ndata(0LL), Ndata_conv(0LL), Ndata_conv0(0LL), TT(0.0) {} + + //Scan_Info::Scan_Info (DP sr, long long int Nf, long long int Ni, long long int Nd, long long int Ndc, long long int Ndc0, long long int t) : + Scan_Info::Scan_Info (DP sr, DP Nf, long long int Ni, long long int Nd, long long int Ndc, long long int Ndc0, double t) : + //sumrule_obtained(sr), Nfull(Nf), Ninadm(Ni), Ndata(Nd), Ndata_conv(Ndc), Ndata_conv0(Ndc0), CPU_ticks(t), CPU_ticks_TOT(t) {} + sumrule_obtained(sr), Nfull(Nf), Ninadm(Ni), Ndata(Nd), Ndata_conv(Ndc), Ndata_conv0(Ndc0), TT(t) {} + + void Scan_Info::Save (const char* outfile_Cstr) + { + ofstream outfile; + + outfile.open(outfile_Cstr); + if (outfile.fail()) JSCerror("Could not open outfile... "); + + //outfile.setf(ios::fixed); + //outfile.setf(ios::showpoint); + outfile.precision(16); + + int TT_hr = int(TT/3600); + int TT_min = int((TT - 3600.0*TT_hr)/60); + + outfile << setw(25) << setprecision(16) << sumrule_obtained << setw(25) << Nfull << setw(16) << Ninadm << setw(16) << Ndata << setw(16) << Ndata_conv << setw(16) << Ndata_conv0 + //<< "\t" << CPU_ticks/CLOCKS_PER_SEC << "\t" << CPU_ticks_TOT/CLOCKS_PER_SEC << endl; + //<< setw(16) << std::fixed << setprecision(3) << TT << endl; + << "\t" << TT_hr << " h " << TT_min << " m " << std::fixed << setprecision(3) << TT - 3600*TT_hr - 60*TT_min << " s" << endl; + //outfile << "sumrule_obtained \t Nfull \t Ninadm \t Ndata \t Ndata_conv \t Ndata_conv0 \t T \t TT."; + outfile << setw(25) << "sumrule_obtained" << setw(25) << "Nfull" << setw(16) << "Ninadm" << setw(16) << "Ndata" << setw(16) << "Ndata_conv" << setw(16) << "Ndata_conv0" << setw(16) << "TT." << endl; + outfile.close(); + + return; + } + + void Scan_Info::Load (const char* infile_Cstr) + { + ifstream infile; + infile.open(infile_Cstr); + if(infile.fail()) { + cout << endl << infile_Cstr << endl; + JSCerror("Could not open input file in Scan_Info::Load."); + } + + int TT_hr, TT_min; + DP TT_sec; + char a; + + //infile >> sumrule_obtained >> Nfull >> Ninadm >> Ndata >> Ndata_conv >> Ndata_conv0 >> CPU_ticks >> CPU_ticks_TOT; + //infile >> sumrule_obtained >> Nfull >> Ninadm >> Ndata >> Ndata_conv >> Ndata_conv0 >> TT; + infile >> sumrule_obtained >> Nfull >> Ninadm >> Ndata >> Ndata_conv >> Ndata_conv0 >> TT_hr >> a >> TT_min >> a >> TT_sec >> a; + + TT = 3600.0 * TT_hr + 60.0* TT_min + TT_sec; + //CPU_ticks_TOT *= CLOCKS_PER_SEC; // correct for factor in Save function + //CPU_ticks = 0; // reset CPU ticks. + + infile.close(); + + return; + } + + std::ostream& operator<< (std::ostream& s, const Scan_Info& info) + { + s.ios::unsetf(ios::scientific); + return s << " sr " << setprecision(14) << info.sumrule_obtained + << "\tNfull " << std::fixed << setprecision(0) << info.Nfull << "\t Ninadm " << info.Ninadm << " Ndata " << info.Ndata + << "\t_conv " << info.Ndata_conv << " _conv0 " << info.Ndata_conv0 + //<< " t " << info.CPU_ticks/CLOCKS_PER_SEC << "s" + //<< " TT " << info.CPU_ticks_TOT/CLOCKS_PER_SEC; + //<< "\tTT " << std::fixed << setprecision(3) << info.TT; + << "\tTT " << int(info.TT/3600) << " h " << int((info.TT - 3600.0 * int(info.TT/3600))/60) << " m " << std::fixed << setprecision(3) << info.TT - 3600.0 * int(info.TT/3600) - 60.0 * int((info.TT - 3600.0 * int(info.TT/3600))/60) << " s"; + } + + +} // namespace JSC diff --git a/src/SCAN/Scan_Thread_Data.cc b/src/SCAN/Scan_Thread_Data.cc new file mode 100644 index 0000000..04fa290 --- /dev/null +++ b/src/SCAN/Scan_Thread_Data.cc @@ -0,0 +1,338 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Scan_Thread_Data.cc + +Purpose: defines all functions for Scan_Thread_Data class. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + Scan_Thread::Scan_Thread() + { + label = ""; + type = -1; + } + + Scan_Thread& Scan_Thread::operator= (const Scan_Thread& RefThread) + { + label = RefThread.label; + type = RefThread.type; + } + + + Scan_Thread_Data::Scan_Thread_Data() + { + } + + Scan_Thread_Data::Scan_Thread_Data(string thrdir_name_ref, bool refine) + { + thrdir_name = thrdir_name_ref; + nthreads_total = Vect (0, nlists); + nthreads_on_disk = Vect (0, nlists); + nthreads_in_memory = Vect (0, nlists); + + lowest_il_with_nthreads_neq_0 = nlists - 1; + + dim = Vect (nlists); // size of the memory vector + label = Vect > (nlists); + type = Vect > (nlists); + + // Give starting values to all: + for (int il = 0; il < nlists; ++il) { + dim[il] = 100; + label[il] = Vect (dim[il]); + type[il] = Vect (dim[il]); + } + + filename = Vect (nlists); + //file = Vect (nlists); + //file_is_open = Vect (nlists); + + for (int il = 0; il < nlists; ++il) { + stringstream filename_strstream; + filename_strstream << thrdir_name << "/" << il << ".thr"; + filename[il] = filename_strstream.str(); + if (!refine) remove(filename[il].c_str()); // the file is deleted to make sure we don't interfere with a previous (failed) computation + //file_is_open[il] = false; + } + if (!refine) { + // remove the nthreads.dat file + stringstream datfile_strstream; + datfile_strstream << thrdir_name << "/nthreads.dat"; + string datfilename = datfile_strstream.str(); + remove(datfilename.c_str()); + } + + } + + Scan_Thread_Data::~Scan_Thread_Data() + { + //for (int il = 0; il < nlists; ++il) + //if (file_is_open[il]) { + // (*file[il]).close(); + // delete file[il]; + //} + } + + bool Scan_Thread_Data::Increase_Memory_Size (int il, int nr_to_add) + { + if (il < 0 || il > nlists) JSCerror("ilist out of bounds in Scan_Thread_Data::Increase_Memory_Size"); + + dim[il] += nr_to_add; + + try { + label[il].Increase_Size (nr_to_add, ""); + type[il].Increase_Size (nr_to_add); + } + + catch (bad_alloc) { + cout << "dim[il] " << dim[il] << "\tnr_to_add " << nr_to_add << endl; + JSCerror("Memory allocation failed in Scan_Thread_Data::Increase_Memory_Size."); + } + + return(true); + } + + void Scan_Thread_Data::Include_Thread (DP abs_data_value_ref, string label_ref, int type_ref) + { + if (abs_data_value_ref <= 0.0) abs_data_value_ref = 1.0e-200; // safety + // Determine which ilist index is to be used: + int il = int(-log(abs_data_value_ref)/logscale); + if (il < 0) il = 0; + if (il >= nlists) il = nlists - 1; + + (*this).Include_Thread (il, label_ref, type_ref); + } + + void Scan_Thread_Data::Include_Thread (int il, string label_ref, int type_ref) + { + //cout << "Calling Include_Threads..." << endl; + + if (il < 0 || il > nlists - 1) JSCerror("il out of range in Scan_Thread_Data::Include_Thread."); + + //cout << "\t\tIncluding thread " << label_ref << "\t" << type_ref << " in list with il = " << il << endl; + + if (il < lowest_il_with_nthreads_neq_0) lowest_il_with_nthreads_neq_0 = il; + + // append to file + //if (!file_is_open[il]) { + //file[il] = new fstream(filename[il].c_str(), fstream::out); + //file_is_open[il] = true; + //} + //*file[il] << label_ref << "\t" << type_ref << endl; + + // Keep in memory for now: + if (nthreads_in_memory[il] > dim[il] - 10) { + (*this).Increase_Memory_Size (il, dim[il]); + } + label[il][nthreads_in_memory[il] ] = label_ref; + type[il][nthreads_in_memory[il] ] = type_ref; + nthreads_in_memory[il]++; + + nthreads_total[il]++; + + // We save the threads to disk if there are sufficiently many: + if (nthreads_in_memory[il] > 1000) { + fstream outfile; + outfile.open(filename[il].c_str(), fstream::out | fstream::app); + for (int it = 0; it < nthreads_in_memory[il]; ++it) + outfile << label[il][it] << "\t" << type[il][it] << endl; + outfile.close(); + nthreads_on_disk[il] += nthreads_in_memory[il]; + + // We then reset these memory buffers + dim[il] = 100; + nthreads_in_memory[il] = 0; + label[il] = Vect (dim[il]); + type[il] = Vect (dim[il]); + } + + //cout << "\t\tDone including thread." << endl; + //char a; + //cin >> a; + //cout << "OK for Include_Threads..." << endl; + + } + + + Vect Scan_Thread_Data::Extract_Next_Scan_Threads () + { + //cout << "Calling Extract_Next_Scan_Threads..." << endl; + + // Returns a vector of threads which are next in line for scanning. + + //cout << "Here 1" << endl; + + int il_used = lowest_il_with_nthreads_neq_0; + Vect next_in_line(nthreads_total[il_used]); + + // Copy the in-memory threads + for (int it = 0; it < nthreads_in_memory[il_used]; ++it) { + next_in_line[it] = Scan_Thread(label[il_used][it], type[il_used][it]); + } + + // Copy the on-disk threads + if (nthreads_on_disk[il_used] > 0) { + ifstream infile; + infile.open(filename[il_used].c_str()); + + string label_read; + int type_read; + for (int in = 0; in < nthreads_on_disk[il_used]; ++in) { + infile >> label_read; + infile >> type_read; + next_in_line[nthreads_in_memory[il_used] + in] = Scan_Thread(label_read, type_read); + } + } + + // The threads in this list are now considered handled. + // Clear memory and remove on-disk file: + nthreads_total[il_used] = 0; + nthreads_on_disk[il_used] = 0; + nthreads_in_memory[il_used] = 0; + label[il_used] = Vect ("", dim[il_used]); + type[il_used] = Vect (dim[il_used]); + remove(filename[il_used].c_str()); + + /* Moved to Include_Thread + // We save the higher-index in-memory threads to files if they are big enough: + for (int il = il_used + 1; il < nlists; ++il) + //if (nthreads_in_memory[il] > 0) { + if (nthreads_in_memory[il] > 1000) { + fstream outfile; + outfile.open(filename[il].c_str(), fstream::out | fstream::app); + for (int it = 0; it < nthreads_in_memory[il]; ++it) + outfile << label[il][it] << "\t" << type[il][it] << endl; + outfile.close(); + nthreads_on_disk[il] += nthreads_in_memory[il]; + + // We then reset these memory buffers + dim[il] = 100; + nthreads_in_memory[il] = 0; + label[il] = Vect (dim[il]); + type[il] = Vect (dim[il]); + } + */ + // Find the next non-empty list: + do { + lowest_il_with_nthreads_neq_0 += 1; + if (lowest_il_with_nthreads_neq_0 == nlists) { + lowest_il_with_nthreads_neq_0 = nlists - 1; + break; + } + } while (nthreads_total[lowest_il_with_nthreads_neq_0] == 0); + + //cout << "Set lowest_il_with_nthreads_neq_0 to " << lowest_il_with_nthreads_neq_0 << endl; + //cin >> a; + + //cout << "OK for Extract_Next_Scan_Threads." << endl; + + return(next_in_line); + } + + Vect Scan_Thread_Data::Extract_Next_Scan_Threads (int min_nr) + { + // Ensures that at least min_nr of threads are returned, if possible + + Vect threads_to_return = Extract_Next_Scan_Threads(); + + while (threads_to_return.size() < min_nr && (*this).lowest_il_with_nthreads_neq_0 < (*this).nlists - 1) { + threads_to_return.Append ((*this).Extract_Next_Scan_Threads()); + } + + return(threads_to_return); + } + + + void Scan_Thread_Data::Flush_to_Disk (int il) + { + if (il < 0 || il > nlists - 1) JSCerror("il out of range in Scan_Thread_Data::Flush_to_Disk."); + + if (nthreads_in_memory[il] > 0) { + fstream outfile; + outfile.open(filename[il].c_str(), fstream::out | fstream::app); + for (int it = 0; it < nthreads_in_memory[il]; ++it) + outfile << label[il][it] << "\t" << type[il][it] << endl; + outfile.close(); + nthreads_on_disk[il] += nthreads_in_memory[il]; + + // We then reset these memory buffers + dim[il] = 100; + nthreads_in_memory[il] = 0; + label[il] = Vect (dim[il]); + type[il] = Vect (dim[il]); + } + } + + void Scan_Thread_Data::Save() + { + // We save the in-memory threads to disk: + for (int il = 0; il < nlists; ++il) (*this).Flush_to_Disk(il); + + ofstream nthreads_outfile; + + stringstream nthreads_outfile_strstream; + nthreads_outfile_strstream << thrdir_name << "/nthreads.dat"; + string nthreads_outfile_str = nthreads_outfile_strstream.str(); + + nthreads_outfile.open(nthreads_outfile_str.c_str()); + if (nthreads_outfile.fail()) JSCerror("Could not open outfile in Scan_Thread_Data::Save... "); + + //cout << "Saving threads: nthreads_tot vector is" << endl; + for (int il = 0; il < nlists; ++il) { + if (nthreads_total[il] != nthreads_in_memory[il] + nthreads_on_disk[il]) + JSCerror("nthreads_total neq _in_memory + _on_disk in Scan_Threads_Data::Save"); + if (nthreads_total[il] > 0) nthreads_outfile << endl << il << "\t" << nthreads_total[il]; + } + + nthreads_outfile.close(); + + return; + } + + void Scan_Thread_Data::Load () + { + ifstream nthreads_infile; + + stringstream nthreads_infile_strstream; + nthreads_infile_strstream << thrdir_name << "/nthreads.dat"; + string nthreads_infile_str = nthreads_infile_strstream.str(); + + nthreads_infile.open(nthreads_infile_str.c_str()); + if (nthreads_infile.fail()) JSCerror("Could not open infile in Scan_Thread_Data::Load... "); + + + // Read the number of elements in each list: + Vect nthreads_read(0, nlists); + int il_read; + + bool min_il_set = false; + while (nthreads_infile.peek() != EOF) { + nthreads_infile >> il_read; + if (!min_il_set) { + lowest_il_with_nthreads_neq_0 = il_read; + min_il_set = true; + } + nthreads_infile >> nthreads_on_disk[il_read]; + nthreads_total[il_read] = nthreads_on_disk[il_read]; + } + + nthreads_infile.close(); + + return; + } + +} // namespace JSC diff --git a/src/SCAN/Scan_Thread_List.cc b/src/SCAN/Scan_Thread_List.cc new file mode 100644 index 0000000..e7b97f8 --- /dev/null +++ b/src/SCAN/Scan_Thread_List.cc @@ -0,0 +1,466 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Scan_Thread_List.cc + +Purpose: defines all functions for Scan_Thread_List class. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + Scan_Thread_List::Scan_Thread_List() + { + dim = 2; + nthreads_tot = 0; + nthreads_done = 0; + //omega = Vect(1.0e+6, dim); + //iK = Vect_INT(0, dim); + //abs_data_value = Vect_DP (0.0, Nentries); + abs_data_value = Vect (0.0, dim); + label = Vect (dim); + type = Vect (dim); + isdone = Vect (false, dim); + } + + Scan_Thread_List::Scan_Thread_List (int Nentries) + { + dim = Nentries; + nthreads_tot = 0; + nthreads_done = 0; + ////omega = Vect_DP(1.0e+6, Nentries); + //omega = Vect(1.0e+6, Nentries); + //iK = Vect_INT(0, Nentries); + //abs_data_value = Vect_DP (0.0, Nentries); + abs_data_value = Vect (0.0, Nentries); + label = Vect (Nentries); + type = Vect (Nentries); + isdone = Vect (false, Nentries); + } + + bool Scan_Thread_List::Increase_Size (int nr_to_add) // resizes the vectors to accommodate up to nr_to_add additional entries + { + dim += nr_to_add; + // Nentries stays the same + + //cout << "Called Increase_Size with nr_to_add = " << nr_to_add << endl; + try { + //omega.Increase_Size (nr_to_add); + //iK.Increase_Size (nr_to_add); + abs_data_value.Increase_Size (nr_to_add); + //cout << "Increasing size of label" << endl; + label.Increase_Size (nr_to_add, ""); + //cout << "Done increasing size of label" << endl; + type.Increase_Size (nr_to_add); + isdone.Increase_Size (nr_to_add, false); + } + + catch (bad_alloc) { + cout << "dim " << dim << "\tnr_to_add " << nr_to_add << endl; + JSCerror("Memory allocation failed in Scan_Thread_List::Increase_Size."); + } + + return(true); + } + + //void Scan_Thread_List::Include_Thread (DP omega_ref, int iK_ref, DP abs_data_value_ref, string label_ref, int type_ref) + void Scan_Thread_List::Include_Thread (DP data_value_ref, string label_ref, int type_ref) + { + if (nthreads_tot > dim - 10) { // Resize the Scan_Thread_List, by doubling its size + //cout << "Resizing threads list" << endl; + (*this).Increase_Size (dim); + //cout << "Done resizing" << endl; + } + + //cout << "Including thread for label " << label_ref << " and type " << type_ref << endl; + + //omega[nthreads] = omega_ref; + //iK[nthreads] = iK_ref; + abs_data_value[nthreads_tot] = abs(data_value_ref); + label[nthreads_tot] = label_ref; + type[nthreads_tot] = type_ref; + isdone[nthreads_tot] = false; + + nthreads_tot++; + + if (nthreads_tot >= dim) { + (*this).Save("Threads_stopped.thr"); + JSCerror("nthreads_tot >= dim in Scan_Thread_List."); + } + + } + + void Scan_Thread_List::Order_in_abs_data_value () + { + // Reorders all entries in decreasing order in abs_data_value + + if (nthreads_tot > 1) { + + Vect_INT index(dim); + for (int i = 0; i < nthreads_tot; ++i) index[i] = i; + + abs_data_value.QuickSort(index, 0, nthreads_tot - 1); + + ////Vect_DP omega_ordered(nthreads); + //Vect omega_ordered(nthreads); + //Vect_INT iK_ordered(nthreads); + //Vect_DP abs_data_value_ordered(nthreads); + Vect abs_data_value_ordered(nthreads_tot); + Vect label_ordered(nthreads_tot); + Vect type_ordered(nthreads_tot); + Vect isdone_ordered(nthreads_tot); + + // Put data in proper order + for (int i = 0; i < nthreads_tot; ++i) { + //omega_ordered[i] = omega[index[nthreads - 1 - i] ]; + //iK_ordered[i] = iK[index[nthreads - 1 - i] ]; + abs_data_value_ordered[i] = abs_data_value[nthreads_tot - 1 - i]; + label_ordered[i] = label[index[nthreads_tot - 1 - i] ]; + type_ordered[i] = type[index[nthreads_tot - 1 - i] ]; + //sector_lowest_raisable_ordered[i] = sector_lowest_raisable[index[nthreads - 1 - i] ]; + isdone_ordered[i] = isdone[index[nthreads_tot - 1 - i] ]; + } + + // Put back in *this object: + for (int i = 0; i < nthreads_tot; ++i) { + //omega[i] = omega_ordered[i]; + //iK[i] = iK_ordered[i]; + abs_data_value[i] = abs_data_value_ordered[i]; + label[i] = label_ordered[i]; + type[i] = type_ordered[i]; + isdone[i] = isdone_ordered[i]; + } // The rest are all simply 0. + + // Done ! + } + } + /* + void Scan_Thread_List::Order_in_omega () + { + // Reorders all entries in increasing order in omega + + if (nthreads > 1) { + + Vect_INT index(0, dim); + for (int i = 0; i < nthreads; ++i) index[i] = i; + + omega.QuickSort(index, 0, nthreads - 1); + + //Vect_DP omega_ordered(nthreads); + Vect omega_ordered(nthreads); + Vect_INT iK_ordered(nthreads); + //Vect_DP abs_data_value_ordered(nthreads); + Vect abs_data_value_ordered(nthreads); + Vect label_ordered(nthreads); + Vect type_ordered(nthreads); + + // Put data in proper order + for (int i = 0; i < nthreads; ++i) { + omega_ordered[i] = omega[i]; + iK_ordered[i] = iK[index[i] ]; + abs_data_value_ordered[i] = abs_data_value[index[i] ]; + label_ordered[i] = label[index[i] ]; + type_ordered[i] = type[index[i] ]; + } + + // Put back in *this object: + for (int i = 0; i < nthreads; ++i) { + omega[i] = omega_ordered[i]; + iK[i] = iK_ordered[i]; + abs_data_value[i] = abs_data_value_ordered[i]; + label[i] = label_ordered[i]; + type[i] = type_ordered[i]; + } // The rest are all simply 0. + + // Done ! + } + } + */ + /* + DP Scan_Thread_List::Highest_abs_data_value_ordered (DP omega_MIN, DP omega_MAX) // returns highest abs_data_value in window omega_MIN to omega_MAX + { + // Assume that the threads have been Order_in_abs_data_value + for (int i = 0; i < nthreads; ++i) + if (omega[i] > omega_MIN && omega[i] < omega_MAX) return(abs_data_value[i]); + + return(0.0); + } + */ + //DP Scan_Thread_List::Highest_abs_data_value (DP omega_MIN, DP omega_MAX) // returns highest abs_data_value in window omega_MIN to omega_MAX + DP Scan_Thread_List::Highest_abs_data_value () // returns highest abs_data_value + { + // Works even if list is not ordered + DP maxdatavalue = 0.0; + for (int i = 0; i < nthreads_tot; ++i) + //if (omega[i] > omega_MIN && omega[i] < omega_MAX && abs_data_value[i] > maxdatavalue) + if (!isdone[i] && (abs_data_value[i] > maxdatavalue)) + maxdatavalue = abs_data_value[i]; + + return(maxdatavalue); + } + + DP Scan_Thread_List::kth_highest_abs_data_value (int k) // returns the k-th highest abs_data_value in window omega_MIN to omega_MAX + { + // Works even if list is not ordered + if (k < 1) JSCerror("Give k > 1 in Scan_Thread_List::kth_Highest_abs_data_value."); + else if (k == 1 || k >= nthreads_tot) // Threads list not long enough, return the top one + //return ((*this).Highest_abs_data_value (-1.0e+10, 1.0e+10)); + return ((*this).Highest_abs_data_value ()); + + Vect topk (0.0, k); + + for (int i = 0; i < nthreads_tot; ++i) + if (!isdone[i] && (abs_data_value[i] > topk[0])) { + topk[0] = abs_data_value[i]; + topk.QuickSort(0, k-1); + } + + return(topk[0]); + } + + bool Scan_Thread_List::Exists_data_value_greater_than (DP value) + { + for (int i = 0; i < nthreads_tot; ++i) + if (abs_data_value[i] > value) return(true); + return(false); + } + + /* + DP Scan_Thread_List::Lowest_omega () // returns lowest omega + { + // Assume that the threads have been Order_in_omega + return(omega[0]); + } + */ + void Scan_Thread_List::Merge (const Scan_Thread_List& reflist) + { + if (nthreads_tot + reflist.nthreads_tot >= dim) // JSCerror("Scan_Threads_List: too big to merge."); + //(*this).Increase_Size (reflist.nthreads); + (*this).Increase_Size (nthreads_tot/10 + reflist.nthreads_tot); + + for (int i = 0; i < reflist.nthreads_tot; ++i) { + //omega[nthreads] = reflist.omega[i]; + //iK[nthreads] = reflist.iK[i]; + abs_data_value[nthreads_tot] = reflist.abs_data_value[i]; + label[nthreads_tot] = reflist.label[i]; + type[nthreads_tot] = reflist.type[i]; + isdone[nthreads_tot] = reflist.isdone[i]; + nthreads_tot++; + } + } + + /* // DEACTIVATED IN ABACUS++G_2 + void Scan_Thread_List::Remove_Threads (int ithread_down, int ithread_up) + { + if (ithread_down < 0 || ithread_up > nthreads) { + JSCerror("Trying to remove inexistent entries in Remove_Threads."); + } + + for (int i = 0; i < nthreads - ithread_up - 1; ++i) { + //omega[ithread_down + i] = omega[ithread_up + i + 1]; + //iK[ithread_down + i] = iK[ithread_up + 1 + i]; + abs_data_value[ithread_down + i] = abs_data_value[ithread_up + i + 1]; + label[ithread_down + i] = label[ithread_up + i + 1]; + type[ithread_down + i] = type[ithread_up + i + 1]; + } + // Zero the other entries: + for (int i = nthreads - ithread_up - 1; i < nthreads; ++i) { + //omega[i] = 1.0e+6; + //iK[i] = 0; + abs_data_value[i] = 0.0;//-1.0e-30; // give a small negative value, so these are put at the very bottom of the list + label[i] = ""; + type[i] = 0; + } + + nthreads -= ithread_up - ithread_down + 1; // removed that many entries. + } + */ + /* // DEACTIVATED IN ABACUS++G_2 + int Scan_Thread_List::Remove_Threads (const Vect& threads_done) + { + // If threads_done[ithread] == true, remove from list + + if (threads_done.size() != nthreads) JSCerror("Wrong size boolean vector in Scan_Thread_List::Remove_Threads."); + + int nr_removed = 0; + + for (int i = 0; i < nthreads; ++i) { + if (!threads_done[i]) { + //omega[i - nr_removed] = omega[i]; + //iK[i - nr_removed] = iK[i]; + abs_data_value[i - nr_removed] = abs_data_value[i]; + label[i - nr_removed] = label[i]; + type[i - nr_removed] = type[i]; + + } + else nr_removed++; + } + // Zero other entries: + for (int i = nthreads - nr_removed; i < nthreads; ++i) { + //omega[i] = 1.0e+6; + //iK[i] = 0; + abs_data_value[i] = 0.0; + label[i] = ""; + type[i] = 0; + } + + nthreads -= nr_removed; + + return(nr_removed); + } + */ + + void Scan_Thread_List::Remove_Done_Threads () + { + // If isdone[ithread] == true, remove from list + + int nr_removed = 0; + + for (int i = 0; i < nthreads_tot; ++i) { + if (!isdone[i]) { + //omega[i - nr_removed] = omega[i]; + //iK[i - nr_removed] = iK[i]; + abs_data_value[i - nr_removed] = abs_data_value[i]; + label[i - nr_removed] = label[i]; + type[i - nr_removed] = type[i]; + isdone[i - nr_removed] = isdone[i]; + } + else nr_removed++; + } + // Zero other entries: + for (int i = nthreads_tot - nr_removed; i < nthreads_tot; ++i) { + //omega[i] = 1.0e+6; + //iK[i] = 0; + abs_data_value[i] = 0.0; + label[i] = ""; + type[i] = 0; + isdone[i] = false; + } + + nthreads_tot -= nr_removed; + if (nthreads_done != nr_removed) { + cout << nthreads_done << "\t" << nr_removed << endl; + JSCerror("Miscount in removing threads during Remove_Done_Threads."); + } + nthreads_done = 0; + + return; + } + + void Scan_Thread_List::Clear () + { + // dim is preserved + + nthreads_tot = 0; + nthreads_done = 0; + ////omega = Vect_DP(0.0, dim); + //omega = Vect (1.0e+6, dim); + //iK = Vect_INT(0, dim); + //abs_data_value = Vect_DP (0.0, dim); + abs_data_value = Vect (0.0, dim); + label = Vect (dim); + type = Vect (dim); + isdone = Vect (false, dim); + } + + void Scan_Thread_List::Save(const char* outfile_Cstr) + { + // We save only the undone threads, so after cleanup. + (*this).Remove_Done_Threads(); + + ofstream outfile; + + outfile.open(outfile_Cstr); + if (outfile.fail()) JSCerror("Could not open outfile... "); + outfile.precision(3); + + if (nthreads_tot > 0) { + //outfile << omega[0] << "\t" << iK[0] << "\t" << abs_data_value[0] << "\t" << label[0] << "\t" << type[0]; + outfile << abs_data_value[0] << "\t" << label[0] << "\t" << type[0]; + for (int i = 1; i < nthreads_tot; ++i) { + //outfile << endl << omega[i] << "\t" << iK[i] << "\t" << abs_data_value[i] << "\t" << label[i] << "\t" << type[i]; + outfile << endl << abs_data_value[i] << "\t" << label[i] << "\t" << type[i]; + } + } + + outfile.close(); + + return; + } + + void Scan_Thread_List::Load (const char* thrfile_Cstr) + { + /* + // Check that the data fits within limits: + struct stat statbuf; + + stat (thrfile_Cstr, &statbuf); + int filesize = statbuf.st_size; + + // Determine the number of entries approximately + // Don't overestimate entry_size: pretend label takes no space... + int entry_size = 2* sizeof(float) + sizeof(int); + + int estimate_nr_entries = filesize/entry_size; + */ + + // Get number of lines in the file: + int nrlines = 0; + ifstream infile1(thrfile_Cstr); + if(infile1.fail()) { + cout << "Could not open file " << thrfile_Cstr << " in Scan_Thread_List::Load." << endl; + JSCerror("Terminating."); + } + string dummy; + while (getline(infile1, dummy)) ++nrlines; + infile1.close(); + + int estimate_nr_entries = JSC::max(10, (11*nrlines)/10); + + if (estimate_nr_entries > dim) { + (*this).Increase_Size (estimate_nr_entries - dim/2); // give dim/2 safety + } + + ifstream infile; + infile.open(thrfile_Cstr); + if(infile.fail()) { + cout << "Could not open file " << thrfile_Cstr << " in Scan_Thread_List::Load." << endl; + JSCerror("Terminating."); + } + + (*this).Clear(); + + DP abs_data_value_DP; + while (infile.peek() != EOF) { + //infile >> omega[nthreads]; + //infile >> iK[nthreads]; + //infile >> abs_data_value[nthreads]; // nasty bug: reading into float bugs when value less or more than float limits + infile >> abs_data_value_DP; // cure: first read into double, then cast to float: + abs_data_value[nthreads_tot] = float(abs_data_value_DP); + infile >> label[nthreads_tot]; + infile >> type[nthreads_tot]; + isdone[nthreads_tot] = false; + nthreads_tot++; + if (nthreads_tot >= dim) { + cout << "file " << thrfile_Cstr << endl << "estimate_nr_entries = " << estimate_nr_entries << "\tnrlines = " << nrlines << endl << "nthreads_tot = " << nthreads_tot << "\tdim = " << dim << endl; + JSCerror("Too many threads in input file."); + } + } + + infile.close(); + + return; + } + +} // namespace JSC diff --git a/src/SCAN/Scan_Thread_Set.cc b/src/SCAN/Scan_Thread_Set.cc new file mode 100644 index 0000000..0c3663e --- /dev/null +++ b/src/SCAN/Scan_Thread_Set.cc @@ -0,0 +1,254 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Scan_Thread_Set.cc + +Purpose: defines all functions for Scan_Thread_Set class. + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + Scan_Thread_Set::Scan_Thread_Set() + { + dim = Vect (nlists); + nthreads_tot = Vect (nlists); + nthreads_done = Vect (nlists); + + label = Vect > (nlists); + type = Vect > (nlists); + isdone = Vect > (nlists); + + + // Give starting values to all: + for (int il = 0; il < nlists; ++il) { + dim[il] = 100; + nthreads_tot[il] = 0; + nthreads_done[il] = 0; + label[il] = Vect (dim[il]); + type[il] = Vect (dim[il]); + isdone[il] = Vect (false, dim[il]); + } + } + + + bool Scan_Thread_Set::Increase_Size (int il, int nr_to_add) // resizes the vectors to accommodate up to nr_to_add additional entries + { + if (il < 0 || il > nlists) JSCerror("ilist out of bounds in Scan_Thread_Set::Increase_Size"); + + dim[il] += nr_to_add; + + try { + label[il].Increase_Size (nr_to_add, ""); + type[il].Increase_Size (nr_to_add); + isdone[il].Increase_Size (nr_to_add, false); + } + + catch (bad_alloc) { + cout << "dim[il] " << dim[il] << "\tnr_to_add " << nr_to_add << endl; + JSCerror("Memory allocation failed in Scan_Thread_Set::Increase_Size."); + } + + return(true); + } + + void Scan_Thread_Set::Include_Thread (DP abs_data_value_ref, string label_ref, int type_ref) + { + // Determine which ilist index is to be used: + int il = int(-log(abs_data_value_ref)/logscale); + if (il < 0) il = 0; + if (il >= nlists) il = nlists - 1; + + if (nthreads_tot[il] > dim[il] - 10) { // Resize the Scan_Thread_Set list, by doubling its size + //cout << "Resizing threads list" << endl; + (*this).Increase_Size (il, dim[il]); + //cout << "Done resizing" << endl; + } + + //cout << "Including thread for label " << label_ref << " and type " << type_ref << endl; + + label[il][nthreads_tot[il] ] = label_ref; + type[il][nthreads_tot[il] ] = type_ref; + isdone[il][nthreads_tot[il] ] = false; + + nthreads_tot[il]++; + + } + + void Scan_Thread_Set::Merge (const Scan_Thread_Set& refset) + { + if (nlists != refset.nlists) JSCerror("nlists don't match in Scan_Thread_Set::Merge"); + + for (int il = 0; il < nlists; ++il) { + + if (nthreads_tot[il] + refset.nthreads_tot[il] >= dim[il]) + (*this).Increase_Size (il, nthreads_tot[il]/10 + refset.nthreads_tot[il]); + + for (int i = 0; i < refset.nthreads_tot[il]; ++i) { + label[il][nthreads_tot[il] ] = refset.label[il][i]; + type[il][nthreads_tot[il] ] = refset.type[il][i]; + isdone[il][nthreads_tot[il] ] = refset.isdone[il][i]; + nthreads_tot[il]++; + } + } + } + + void Scan_Thread_Set::Remove_Done_Threads (int il) + { + // If isdone[ithread] == true, remove from list + + int nr_removed = 0; + + for (int i = 0; i < nthreads_tot[il]; ++i) { + if (!isdone[il][i]) { + label[il][i - nr_removed] = label[il][i]; + type[il][i - nr_removed] = type[il][i]; + isdone[il][i - nr_removed] = isdone[il][i]; + } + else nr_removed++; + } + // Zero other entries: + for (int i = nthreads_tot[il] - nr_removed; i < nthreads_tot[il]; ++i) { + label[il][i] = ""; + type[il][i] = 0; + isdone[il][i] = false; + } + + nthreads_tot[il] -= nr_removed; + if (nthreads_done[il] != nr_removed) { + cout << nthreads_done[il] << "\t" << nr_removed << endl; + JSCerror("Miscount in removing threads during Scan_Thread_Set::Remove_Done_Threads."); + } + nthreads_done[il] = 0; + + return; + } + + void Scan_Thread_Set::Remove_Done_Threads () + { + // If isdone[ithread] == true, remove from list + + for (int il = 0; il < nlists; ++il) + (*this).Remove_Done_Threads(il); + + return; + } + + void Scan_Thread_Set::Clear () + { + for (int il = 0; il < nlists; ++il) { + + nthreads_tot[il] = 0; + nthreads_done[il] = 0; + label[il] = Vect (dim[il]); + type[il] = Vect (dim[il]); + isdone[il] = Vect (false, dim[il]); + } + } + + void Scan_Thread_Set::Save(const char* outfile_Cstr) + { + // We save only the undone threads, so after cleanup. + (*this).Remove_Done_Threads(); + + ofstream outfile; + + outfile.open(outfile_Cstr); + if (outfile.fail()) JSCerror("Could not open outfile... "); + outfile.precision(3); + + //cout << "Saving threads: nthreads_tot vector is" << endl; + //for (int il = 0; il < nlists; ++il) + //if (nthreads_tot[il] > 0) cout << il << "\t" << nthreads_tot[il] << "\t"; + //cout << endl; + + bool started = false; + for (int il = 0; il < nlists; ++il) { + + if (nthreads_tot[il] > 0) { + if (started) outfile << endl; + outfile << il << "\t" << label[il][0] << "\t" << type[il][0]; + started = true; + for (int i = 1; i < nthreads_tot[il]; ++i) { + outfile << endl << il << "\t" << label[il][i] << "\t" << type[il][i]; + } + } + } + + outfile.close(); + + return; + } + + void Scan_Thread_Set::Load (const char* thrfile_Cstr) + { + ifstream infile; + infile.open(thrfile_Cstr); + if(infile.fail()) { + cout << "Could not open file " << thrfile_Cstr << " in Scan_Thread_Set::Load." << endl; + JSCerror("Terminating."); + } + + (*this).Clear(); + + // First count the number of elements in each list: + Vect nthreads_read(0, nlists); + int il_read; + string label_read; + int type_read; + + while (infile.peek() != EOF) { + infile >> il_read; + infile >> label_read; + infile >> type_read; + nthreads_read[il_read]++; + } + + infile.close(); + + //cout << "nthreads_read vector: " << endl; + //for (int il = 0; il < nlists; ++il) + //if (nthreads_read[il] > 0) cout << il << "\t" << nthreads_read[il] << "\t"; + //cout << endl; + + // Now allocate the proper sizes: + for (int il = 0; il < nlists; ++il) { + (*this).Increase_Size (il, JSC::max(0, nthreads_read[il] - dim[il] + 10)); + nthreads_read[il] = 0; + } + + // Read threads data in: + ifstream infile2; + infile2.open(thrfile_Cstr); + infile.seekg(0); + while (infile2.peek() != EOF) { + infile2 >> il_read; + infile2 >> label[il_read][nthreads_read[il_read] ]; + infile2 >> type[il_read][nthreads_read[il_read] ]; + //isdone[il_read][nthreads_read[il_read] ] = false; // no need + nthreads_tot[il_read]++; + nthreads_read[il_read]++; + } + + infile2.close(); + + //cout << "Loading threads: nthreads_tot vector is" << endl; + //for (int il = 0; il < nlists; ++il) + //if (nthreads_tot[il] > 0) cout << il << "\t" << nthreads_tot[il] << "\t"; + //cout << endl; + + return; + } + +} // namespace JSC diff --git a/src/TBA/Root_Density.cc b/src/TBA/Root_Density.cc new file mode 100644 index 0000000..4b81c20 --- /dev/null +++ b/src/TBA/Root_Density.cc @@ -0,0 +1,462 @@ +/********************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) 2008. + +----------------------------------------------------------- + +File: Root_Density.cc + +Purpose: defines Root_Density & Root_Density_Set, and their member functions + +Last modified: 27/09/08 + +Chronology: +08 10 28: created + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + +/************************************/ +/* +struct Root_Density { + + int Npts; // how many points are used to describe each function + DP lambdamax; // what the largest rapidity is + Vect_DP lambda; // rapidity vector + Vect_DP dlambda; // differential element + Vect_DP value; // the root density itself + Vect_DP prev_value; // results of previous iteration + DP diff; // relative differences with previous iteration + bool value_infty_set; // boolean, true if asymptotic value set + DP value_infty; // asymptotic value, computed analytically + + Root_Density (); + Root_Density (int Npts_ref, DP lambdamax_ref); + + Root_Density& operator= (const Root_Density& RefDensity); + + DP Return_Value (DP lambda_ref); // evaluates the function for any argument using linear interpolation + DP Set_Asymptotics (DP value_infty_ref); // sets value for lambda >= lambdamax + + Root_Density Compress_and_Match_Densities (DP comp_factor); // returns a Root_Density with fewer points +}; +*/ + +Root_Density::Root_Density () + : Npts(1), lambdamax(0.0), lambda(Vect_DP(0.0, Npts)), dlambda(Vect_DP(0.0, Npts)), value(Vect_DP(0.0, Npts)), + prev_value(Vect_DP(0.0, Npts)), diff(1.0e+6), value_infty_set(false), value_infty(0.0) +{ +} + +Root_Density::Root_Density (int Npts_ref, DP lambdamax_ref) + : Npts(Npts_ref), lambdamax(lambdamax_ref), lambda(Vect_DP(Npts)), dlambda(Vect_DP(0.0, Npts)), value(Vect_DP(0.0, Npts)), + prev_value(Vect_DP(0.0, Npts)), diff(1.0e+6), value_infty_set(false), value_infty(0.0) +{ + for (int i = 0; i < value.size(); ++i) { + lambda[i] = lambdamax * (-(Npts - 1.0) + 2*i)/Npts; + dlambda[i] = 2.0 * lambdamax/Npts; + } +} + +Root_Density& Root_Density::operator= (const Root_Density& RefDensity) +{ + if (this != &RefDensity) { + Npts = RefDensity.Npts; + lambdamax = RefDensity.lambdamax; + lambda = RefDensity.lambda; + dlambda = RefDensity.dlambda; + value = RefDensity.value; + prev_value = RefDensity.prev_value; + diff = RefDensity.diff; + value_infty_set = RefDensity.value_infty_set; + value_infty = RefDensity.value_infty; + + } + return(*this); +} + +DP Root_Density::Return_Value (DP lambda_ref) +{ + // This function returns a value for epsilon at any real lambda + // using simple linear interpolation. + // Degree 3 polynomical also programmed in, but commented out: no improvement. + + DP answer = 0.0; + if (fabs(lambda_ref) >= fabs(lambda[0])) { + if (value_infty_set) answer = value_infty; + else JSCerror("Need to set asymptotics of Root_Density !"); + } + + else { // try to find the i such that lambda[i] <= lambda_ref < lambda[i+1] + + int index = (Npts - 1)/2; + int indexstep = (Npts - 1)/4 + 1; + + while (indexstep >= 1) { + + if ( // if is "lower": we go up + lambda_ref >= lambda[index + 1]) { + index += indexstep; + } + + else if ( // if is "higher" or equal: we go down + lambda[index] > lambda_ref) { + index -= indexstep; + } + + index = JSC::max(0, index); + index = JSC::min(Npts - 2, index); + + if (indexstep == 1) indexstep--; + else indexstep = (indexstep + 1)/2; + + } // while ... + + if (index < 0 || index >= Npts || lambda[index] > lambda_ref || lambda[index + 1] < lambda_ref) { + cout << "Seeking index: " << index << "\t" << lambda[index] << "\t <=? " << lambda_ref << "\t Npts - 3) + answer = ((value[index] * (lambda[index+1] - lambda_ref) + + value[index + 1] * (lambda_ref - lambda[index]))/(lambda[index+1] - lambda[index])); + /* + else { + // Better: if possible, fit to polynomial going through 4 closest points + Vect_DP xa (4); + Vect_DP ya (4); + DP dy; + xa[0] = lambda[index - 1]; xa[1] = lambda[index]; xa[2] = lambda[index + 1]; xa[3] = lambda[index + 2]; + ya[0] = value[index - 1]; ya[1] = value[index]; ya[2] = value[index + 1]; ya[3] = value[index + 2]; + polint (xa, ya, lambda_ref, answer, dy); // sets answer to value at lambda_ref + } + */ + } + + return(answer); +} + +void Root_Density::Set_Asymptotics (DP value_infty_ref) +{ + value_infty = value_infty_ref; + value_infty_set = true; +} + +Root_Density Root_Density::Compress_and_Match_Densities (DP comp_factor) +{ + // Returns a 'compressed' version of the density, using 1/comp_factor as many points. + + // PROBLEM: this implementation can lead to numerical instabilities. + + //Root_Density compressed_density(Npts/comp_factor, lambdamax); + + // Rather: use this implementation: + int Npts_used = int(2.0 * lambdamax/(dlambda[0] * comp_factor)); + + Root_Density compressed_density(Npts_used, lambdamax); + + compressed_density.Set_Asymptotics (value_infty); + + for (int i = 0; i < compressed_density.Npts; ++i) + compressed_density.value[i] = (*this).Return_Value (compressed_density.lambda[i]); + + return(compressed_density); +} + +void Root_Density::Save (const char* outfile_Cstr) +{ + ofstream outfile; + outfile.open(outfile_Cstr); + outfile.precision(16); + + for (int i = 0; i < Npts; ++i) { + if (i > 0) outfile << endl; + outfile << setw(20) << lambda[i] << "\t" << setw(20) << value[i]; + } + + outfile.close(); +} + + +/************************************/ +/* +struct Root_Density_Set { + + int ntypes; + Vect epsilon; + int Npts_total; // sum of all Npts of epsilon's + DP diff; // sum of diff's of the epsilon's + + Root_Density_Set (); + Root_Density_Set (int ntypes_ref, int Npts_ref, DP lambdamax_ref); + Root_Density_Set (int ntypes_ref, Vect_INT Npts_ref, Vect_DP lambdamax_ref); + + Root_Density_Set& operator= (const Root_Density_Set& RefSet); + + void Insert_new_function (DP asymptotic_value); + void Extend_limits (Vect need_to_extend_limit); + void Insert_new_points (Vect > need_new_point_around); + + DP Return_Value (int n_ref, DP lambda_ref); // returns a value, no matter what. + + Root_Density_Set Return_Compressed_and_Matched_Set (DP comp_factor); + void Match_Densities (Root_Density_Set& RefSet); + + void Save (const char* outfile_Cstr); +}; +*/ +Root_Density_Set::Root_Density_Set () : ntypes(1), epsilon(Vect (ntypes)), Npts_total(0), diff(1.0e+6) +{ +} + +Root_Density_Set::Root_Density_Set (int ntypes_ref, int Npts_ref, DP lambdamax_ref) + : ntypes(ntypes_ref), epsilon(Vect (ntypes_ref)), Npts_total(ntypes_ref * Npts_ref), diff(1.0e+6) +{ + for (int n = 0; n < ntypes; ++n) epsilon[n] = Root_Density(Npts_ref, lambdamax_ref); +} + +Root_Density_Set::Root_Density_Set (int ntypes_ref, Vect_INT Npts_ref, Vect_DP lambdamax_ref) + : ntypes(ntypes_ref), epsilon(Vect (ntypes_ref)), Npts_total(Npts_ref.sum()), diff(1.0e+6) +{ + if (Npts_ref.size() != ntypes_ref || lambdamax_ref.size() != ntypes_ref) JSCerror("Wrong vector sizes in Root_Density_Set."); + for (int n = 0; n < ntypes; ++n) epsilon[n] = Root_Density(Npts_ref[n], lambdamax_ref[n]); +} + +Root_Density_Set& Root_Density_Set::operator= (const Root_Density_Set& RefSet) +{ + if (this != &RefSet) { + ntypes = RefSet.ntypes; + epsilon = RefSet.epsilon; + Npts_total = RefSet.Npts_total; + diff = RefSet.diff; + } + return(*this); +} + +void Root_Density_Set::Insert_new_function (DP asymptotic_value) +{ + // This function extends a set by adding one epsilon_n function on top + + Root_Density_Set Updated_Set (ntypes + 1, 10, 10.0); // last two parameters are meaningless + for (int n = 0; n < ntypes; ++n) Updated_Set.epsilon[n] = epsilon[n]; + + //Updated_Set.epsilon[ntypes] = Root_Density (epsilon[ntypes - 1].Npts, epsilon[ntypes - 1].lambdamax); + Updated_Set.epsilon[ntypes] = Root_Density (50, epsilon[ntypes - 1].lambdamax); + Updated_Set.epsilon[ntypes].Set_Asymptotics (asymptotic_value); + + for (int i = 0; i < Updated_Set.epsilon[ntypes].Npts; ++i) + Updated_Set.epsilon[ntypes].value[i] = Updated_Set.epsilon[ntypes].value_infty; + + ntypes = Updated_Set.ntypes; + epsilon = Updated_Set.epsilon; + Npts_total+= Updated_Set.epsilon[ntypes - 1].Npts; +} + +void Root_Density_Set::Extend_limits (Vect need_to_extend_limit) +{ + // Extend the limits of integration at each level, according to boolean + + // The function extends the limits by 10% on both sides, putting the + // extra values to value_infty. + + if (need_to_extend_limit.size() != epsilon.size()) JSCerror("Wrong size need_to_extend_limit boolean in Extend_limits."); + + Vect_INT nr_new_points_needed(0, ntypes); + int total_nr_new_points_added = 0; + DP dlambda_used = 0.0; + for (int n = 0; n < ntypes; ++n) { + if (need_to_extend_limit[n]) { + + Root_Density epsilon_n_before_update = epsilon[n]; + + // Determine the dlambda to be used: + dlambda_used = epsilon[n].dlambda[0]; + + // How many new points do we add ? Say 5\% on each side: + nr_new_points_needed[n] = JSC::max(1, epsilon[n].Npts/20); + + epsilon[n] = Root_Density(epsilon_n_before_update.Npts + 2* nr_new_points_needed[n], epsilon_n_before_update.lambdamax + nr_new_points_needed[n] * dlambda_used); + epsilon[n].Set_Asymptotics(epsilon_n_before_update.value_infty); + + for (int i = 0; i < nr_new_points_needed[n]; ++i) { + epsilon[n].lambda[i] = epsilon_n_before_update.lambda[0] - (nr_new_points_needed[n] - i) * dlambda_used; + epsilon[n].dlambda[i] = dlambda_used; + epsilon[n].value[i] = epsilon_n_before_update.value_infty; + } + + for (int i = 0; i < epsilon_n_before_update.Npts; ++i) { + epsilon[n].lambda[i + nr_new_points_needed[n] ] = epsilon_n_before_update.lambda[i]; + epsilon[n].dlambda[i + nr_new_points_needed[n] ] = epsilon_n_before_update.dlambda[i]; + epsilon[n].value[i + nr_new_points_needed[n] ] = epsilon_n_before_update.value[i]; + } + + for (int i = 0; i < nr_new_points_needed[n]; ++i) { + epsilon[n].lambda[i + epsilon_n_before_update.Npts + nr_new_points_needed[n] ] + = epsilon_n_before_update.lambda[epsilon_n_before_update.Npts - 1] + (i+1.0) * dlambda_used; + epsilon[n].dlambda[i + epsilon_n_before_update.Npts + nr_new_points_needed[n] ] = dlambda_used; + epsilon[n].value[i + epsilon_n_before_update.Npts + nr_new_points_needed[n] ] = epsilon_n_before_update.value_infty; + } + + total_nr_new_points_added += 2 * nr_new_points_needed[n]; + + //cout << "Extending limits at level " << n << " with " << nr_new_points_needed[n] << " points on each side to " << epsilon[n].lambdamax << endl; + + } // if (need + } // for n + + Npts_total += total_nr_new_points_added; + + // Done ! + + return; + +} + +void Root_Density_Set::Insert_new_points (Vect > need_new_point_around) +{ + // need_new_point_around specifies whether a new point needs to be inserted around existing points. + + // Count the number of new points needed per type: + Vect_INT nr_new_points_needed(0, ntypes); + int total_nr_new_points_needed = 0; + for (int n = 0; n < ntypes; ++n) { + if (need_new_point_around[n].size() != epsilon[n].Npts) JSCerror("Wrong size need_new_point_around boolean in Insert_new_points."); + for (int i = 0; i < epsilon[n].Npts; ++i) + if (need_new_point_around[n][i]) nr_new_points_needed[n]++; + total_nr_new_points_needed += nr_new_points_needed[n]; + } + /* + // Simplistic version: always keep equidistant points + for (int n = 0; n < ntypes; ++n) { + Root_Density epsilon_n_before_update = epsilon[n]; + epsilon[n] = Root_Density(epsilon_n_before_update.Npts + nr_new_points_needed[n], epsilon_n_before_update.lambdamax); + epsilon[n].Set_Asymptotics(epsilon_n_before_update.value_infty); + for (int i = 0; i < epsilon[n].Npts; ++i) + epsilon[n].value[i] = epsilon_n_before_update.Return_Value(epsilon[n].lambda[i]); + } + */ + + // Working version using non-equispaced points + // Now update all data via interpolation: + for (int n = 0; n < ntypes; ++n) { + Root_Density epsilon_n_before_update = epsilon[n]; + epsilon[n] = Root_Density(epsilon_n_before_update.Npts + nr_new_points_needed[n], epsilon_n_before_update.lambdamax); + epsilon[n].Set_Asymptotics(epsilon_n_before_update.value_infty); + //cout << "Check: " << epsilon[n].Npts << " " << epsilon_n_before_update.Npts << endl; + int nr_pts_added_n = 0; + for (int i = 0; i < epsilon_n_before_update.Npts; ++i) { + if (!need_new_point_around[n][i]) { + epsilon[n].lambda[i + nr_pts_added_n] = epsilon_n_before_update.lambda[i]; + epsilon[n].dlambda[i + nr_pts_added_n] = epsilon_n_before_update.dlambda[i]; + epsilon[n].value[i + nr_pts_added_n] = epsilon_n_before_update.value[i]; + } + else if (need_new_point_around[n][i]) { + epsilon[n].lambda[i + nr_pts_added_n] = epsilon_n_before_update.lambda[i] - 0.25 * epsilon_n_before_update.dlambda[i]; + epsilon[n].dlambda[i + nr_pts_added_n] = 0.5 * epsilon_n_before_update.dlambda[i]; + epsilon[n].value[i + nr_pts_added_n] = epsilon_n_before_update.Return_Value(epsilon[n].lambda[i + nr_pts_added_n]); + nr_pts_added_n++; + epsilon[n].lambda[i + nr_pts_added_n] = epsilon_n_before_update.lambda[i] + 0.25 * epsilon_n_before_update.dlambda[i]; + epsilon[n].dlambda[i + nr_pts_added_n] = 0.5 * epsilon_n_before_update.dlambda[i]; + epsilon[n].value[i + nr_pts_added_n] = epsilon_n_before_update.Return_Value(epsilon[n].lambda[i + nr_pts_added_n]); + } + } + if (nr_pts_added_n != nr_new_points_needed[n]) { + cout << nr_pts_added_n << "\t" << nr_new_points_needed[n] << endl; + JSCerror("Wrong counting of new points in Insert_new_points."); + } + + // Check: + //for (int i = 0; i < epsilon[n].Npts - 1; ++i) + //if (fabs(epsilon[n].lambda[i] + 0.5 *(epsilon[n].dlambda[i] + epsilon[n].dlambda[i+1]) - epsilon[n].lambda[i+1]) > 1.0e-13) + //{ + // cout << "Error at level " << n << "\ti " << i << "\t" << epsilon[n].lambda[i] << "\t" << epsilon[n].dlambda[i] + // << "\t" << epsilon[n].lambda[i+1] << "\t" << epsilon[n].dlambda[i+1] + // << "\t" << epsilon[n].lambda[i] + 0.5 *(epsilon[n].dlambda[i] + epsilon[n].dlambda[i+1]) - epsilon[n].lambda[i+1] << endl; + // JSCerror("..."); + //} + + } // for n + + + //cout << "need_new_pt_above " << need_new_point_above[0] << endl << endl; + //cout << "epsilon[0].lambda = " << epsilon[0].lambda << endl << endl; + //cout << "epsilon[0].dlambda = " << epsilon[0].dlambda << endl << endl; + //cout << "epsilon[0].value = " << epsilon[0].value << endl << endl; + + Npts_total += total_nr_new_points_needed; + + // Done ! + + return; +} + +DP Root_Density_Set::Return_Value (int n_ref, DP lambda_ref) +{ + // Returns a value, no matter what ! + + if (n_ref < ntypes) return(epsilon[n_ref].Return_Value(lambda_ref)); + + else // assume asymptotic form of epsilon, proportional to n + return(epsilon[ntypes - 1].Return_Value(lambda_ref) * n_ref/(ntypes - 1.0)); + +} + +Root_Density_Set Root_Density_Set::Return_Compressed_and_Matched_Set (DP comp_factor) +{ // Returns a set with 1/comp_factor as many points at each level + + if (comp_factor >= 2.0) + JSCerror("Compression factor too large in Return_Compressed_and_Matched_Set, numerical instability will occur."); + + Vect_INT nrpts_comp (ntypes); + Vect_DP lambdamax_comp (ntypes); + for (int n = 0; n < ntypes; ++n) { + nrpts_comp[n] = int(2.0 * epsilon[n].lambdamax/(epsilon[n].dlambda[0] * comp_factor)); + lambdamax_comp[n] = epsilon[n].lambdamax; + } + + Root_Density_Set Compressed_and_Matched_Set (ntypes, nrpts_comp, lambdamax_comp); + + for (int n = 0; n < ntypes; ++n) + Compressed_and_Matched_Set.epsilon[n] = (*this).epsilon[n].Compress_and_Match_Densities (comp_factor); + + return(Compressed_and_Matched_Set); +} + +void Root_Density_Set::Match_Densities (Root_Density_Set& RefSet) +{ // matched densities to those in RefSet + + for (int n = 0; n < ntypes; ++n) + for (int i = 0; i < epsilon[n].Npts; ++i) + epsilon[n].value[i] = RefSet.epsilon[n].Return_Value(epsilon[n].lambda[i]); +} + +void Root_Density_Set::Save (const char* outfile_Cstr) +{ + ofstream outfile; + outfile.open(outfile_Cstr); + outfile.precision(16); + + // Determine what the maximal nr of pts is: + int Npts_n_max = 0; + for (int n = 0; n < ntypes; ++n) Npts_n_max = JSC::max(Npts_n_max, epsilon[n].Npts); + + for (int i = 0; i < Npts_n_max; ++i) { + if (i > 0) outfile << endl; + for (int n = 0; n < ntypes; ++n) (i < epsilon[n].Npts) ? + (outfile << epsilon[n].lambda[i] << "\t" << epsilon[n].value[i] << "\t") + : (outfile << 0 << "\t" << 0 << "\t"); + } + + outfile.close(); +} + + +} // namespace JSC diff --git a/src/TBA/TBA_2CBG.cc b/src/TBA/TBA_2CBG.cc new file mode 100644 index 0000000..b7c6ba6 --- /dev/null +++ b/src/TBA/TBA_2CBG.cc @@ -0,0 +1,1738 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: TBA_2CBG.cc + +Purpose: solves the TBA equations for the 2-component Bose gas + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + +/********************* 2CBG specific *******************/ + + +DP Asymptotic_2CBG_epsilon (int n, DP Omega, DP kBT) +{ + return(2.0 * Omega * n + kBT * log(pow((1.0 - exp(-2.0 * (n + 1.0) * Omega/kBT)) + /(1.0 - exp(-2.0 * Omega/kBT)), 2.0) - exp(-2.0 * n * Omega/kBT))); +} + +void Set_2CBG_Asymptotics (Root_Density_Set& TBA_Set, DP mu, DP Omega, DP kBT) +{ + //DP epsilon_infty = 0.0; + TBA_Set.epsilon[0].Set_Asymptotics (pow(TBA_Set.epsilon[0].lambdamax, 2.0) - mu - Omega); + for (int n = 1; n < TBA_Set.ntypes; ++n) { + //epsilon_infty = 2.0 * Omega * n + kBT * + //log(pow((1.0 - exp(-2.0 * (n + 1.0) * Omega/kBT))/(1.0 - exp(-2.0 * Omega/kBT)), 2.0) - exp(-2.0 * n * Omega/kBT)); + //TBA_Set.epsilon[n].Set_Asymptotics (epsilon_infty); + TBA_Set.epsilon[n].Set_Asymptotics (Asymptotic_2CBG_epsilon(n, Omega, kBT)); + //cout << "Set asymptotics of " << n << " to " << setprecision(16) << TBA_Set.epsilon[n].value_infty + // << "\t" << 2.0 * Omega * n + 2.0 * kBT * log((1.0 - exp(-2.0 * (n + 1) * Omega/kBT))/(1.0 - exp(-2.0 * Omega/kBT))) - TBA_Set.epsilon[n].value_infty << endl; + } +} + +void Set_2CBG_deps_dchempot_Asymptotics (int option, const Root_Density_Set& Set, Root_Density_Set& DSet, DP mu, DP Omega, DP kBT) +{ + // option == 0: deps/dmu + // option == 1: deps/dOmega + + //DSet.epsilon[0].Set_Asymptotics (-1.0); + DP zeroasymptote = -1.0; + DP em2OoT = exp(-2.0 * Omega/kBT); + for (int n = 1; n < DSet.ntypes; ++n) { + if (option == 0) DSet.epsilon[n].Set_Asymptotics (0.0); + else if (option == 1) + DSet.epsilon[n].Set_Asymptotics (2.0 * (1.0 - pow(em2OoT, n+1.0)) + * (n * (1.0 - pow(em2OoT, n+2.0)) - (n + 2.0) * em2OoT * (1.0 - pow(em2OoT, DP(n)))) + /((1.0 - em2OoT) * (1.0 - pow(em2OoT, DP(n))) * (1.0 - pow(em2OoT, n + 2.0)))); + //cout << "Set asymptotics for option " << option << " at level " << n << " to " << DSet.epsilon[n].value_infty << endl; + zeroasymptote += DSet.epsilon[n].value_infty * exp(-Set.epsilon[n].value_infty/kBT)/(1.0 + exp(-Set.epsilon[n].value_infty/kBT)); + } + // For n > nmax sum in RHS of BE for epsilon, assuming epsilon_n = epsilon_n^\infty in those cases: + // Remember: nmax in notes is Set.ntypes - 1. + zeroasymptote -= option == 0 ? 0.0 : 2.0 * ((Set.ntypes + 1.0) * exp(-2.0 * (Set.ntypes + 1.0) * Omega/kBT)/(1.0 - exp(-2.0 * (Set.ntypes + 1.0) * Omega/kBT)) + - Set.ntypes * exp(-2.0 * Set.ntypes * Omega/kBT)/(1.0 - exp(-2.0 * Set.ntypes * Omega/kBT))); + + //cout << "Set asymptotics for option " << option << " at level 0 to " << zeroasymptote << endl; + + DSet.epsilon[0].Set_Asymptotics (zeroasymptote); + + return; +} + +void Initiate_2CBG_TBA_Functions (Root_Density_Set& TBA_Set, DP mu, DP Omega) +{ + for (int i = 0; i < TBA_Set.epsilon[0].Npts; ++i) { + TBA_Set.epsilon[0].value[i] = TBA_Set.epsilon[0].lambda[i] * TBA_Set.epsilon[0].lambda[i] - mu - Omega; + TBA_Set.epsilon[0].prev_value[i] = TBA_Set.epsilon[0].value[i]; + } + for (int n = 1; n < TBA_Set.ntypes; ++n) { + for (int i = 0; i < TBA_Set.epsilon[n].Npts; ++i) + TBA_Set.epsilon[n].value[i] = TBA_Set.epsilon[n].value_infty; + } +} + +void Initiate_2CBG_deps_dchempot_Functions (Root_Density_Set& DSet) +{ + for (int n = 0; n < DSet.ntypes; ++n) { + for (int i = 0; i < DSet.epsilon[n].Npts; ++i) + DSet.epsilon[n].value[i] = DSet.epsilon[n].value_infty; + } +} + +void Iterate_2CBG_TBAE (Root_Density_Set& Set, Vect >& a_n_dlambda, Vect >& fmin_dlambda, + Vect >& fplus_dlambda, DP c_int, DP mu, DP Omega, DP kBT) +{ + // Produces a new Root_Density_Set from a previous iteration. + // Does NOT add types or change Npts, lambdamax values. + + //DP oneoverc = 1.0/c_int; + //DP twoovernc = 2.0/c_int; + + // First define some useful functions: + Vect Tln1plusemineps(Set.ntypes); + Vect_DP Tln1pluseminepsinfty(Set.ntypes); + + for (int n = 0; n < Set.ntypes; ++n) { + + Tln1plusemineps[n] = Vect_DP (0.0, Set.epsilon[n].Npts); + + for (int i = 0; i < Set.epsilon[n].Npts; ++i) { + Tln1plusemineps[n][i] = Set.epsilon[n].value[i] > 0.0 ? + kBT * (Set.epsilon[n].value[i] < 24.0 * kBT ? log(1.0 + exp(-Set.epsilon[n].value[i]/kBT)) : exp(-Set.epsilon[n].value[i]/kBT)) + : + -Set.epsilon[n].value[i] + kBT * (-Set.epsilon[n].value[i] < 24.0 * kBT ? log (1.0 + exp(Set.epsilon[n].value[i]/kBT)) : exp(Set.epsilon[n].value[i]/kBT)); + // Keep previous rapidities: + Set.epsilon[n].prev_value[i] = Set.epsilon[n].value[i]; + } + Tln1pluseminepsinfty[n] = kBT * (Set.epsilon[n].value_infty < 24.0 * kBT ? log(1.0 + exp(-Set.epsilon[n].value_infty/kBT)) : exp(-Set.epsilon[n].value_infty/kBT)); + //Tln1pluseminepsinfty[n] = kBT * log(1.0 + exp(-Set.epsilon[n].value_infty/kBT)); + //cout << "Check Tln1pluseminepsinfty: n " << n << " " << Tln1pluseminepsinfty[n] << " " << -kBT * log(1.0 - pow(sinh(Omega/kBT)/sinh((n + 1) * Omega/kBT), 2.0)) << endl; + } + + // Now do the necessary convolutions for epsilon == epsilon[0]. + // For each value of lambda, do the convolutions: + // Careful: the lambda's used for lambda (index i) are those of epsilon[0], the lambda' (index j) are for epsilon[n] !! + Vect a_n_Tln_conv(Set.ntypes); + for (int n = 0; n < Set.ntypes; ++n) { + a_n_Tln_conv[n] = Vect_DP (0.0, Set.epsilon[0].Npts); + Vect_DP f(0.0, Set.epsilon[n].Npts); + + for (int i = 0; i < Set.epsilon[0].Npts; ++i) { + a_n_Tln_conv[n][i] = 0.0; + + for (int j = 0; j < Set.epsilon[n].Npts; ++j) a_n_Tln_conv[n][i] += Tln1plusemineps[n][j] * a_n_dlambda[i][n][j]; + // Add alpha curvature terms: VERY COSTLY + /* + for (int j = 1; j < Set.epsilon[n].Npts - 1; ++j) + a_n_Tln_conv[n][i] += (1.0/12.0) * pow(Set.epsilon[n].dlambda[j], 3.0) + * ((Tln1plusemineps[n][j+1] * a_n_dlambda[i][n][j+1] - Tln1plusemineps[n][j] * a_n_dlambda[i][n][j])/(Set.epsilon[n].lambda[j+1] - Set.epsilon[n].lambda[j]) + - (Tln1plusemineps[n][j] * a_n_dlambda[i][n][j] - Tln1plusemineps[n][j-1] * a_n_dlambda[i][n][j-1])/(Set.epsilon[n].lambda[j] - Set.epsilon[n].lambda[j-1])) + /(Set.epsilon[n].lambda[j+1] - Set.epsilon[n].lambda[j-1]); + */ + } // for (int i ... + + } // for (int n... We now have all the a_n * Tln... at our disposal. + + // For n > nmax sum in RHS of BE for epsilon, assuming epsilon_n = epsilon_n^\infty in those cases: + // Remember: nmax = Set.ntypes - 1 + DP Smaxsum = kBT * log((1.0 - exp(-2.0 * (Set.ntypes + 1.0) * Omega/kBT))/(1.0 - exp(-2.0 * Set.ntypes * Omega/kBT))); + /* + // Check of convolutions, for lambda around 0: + Vect_DP a_n_Tln_conv_0_integ (0.0, Set.ntypes); + Vect_DP integ_target(0.0, Set.ntypes); + //for (int n = 1; n < Set.ntypes; ++n) { + for (int n = Set.ntypes - 1; n < Set.ntypes; ++n) { + for (int j = 0; j < Set.epsilon[n].Npts; ++j) a_n_Tln_conv_0_integ[n] += Tln1plusemineps[n][j] * a_n_dlambda[Set.epsilon[0].Npts/2][n][j]; + // Add asymptotic parts: not necessary + a_n_Tln_conv_0_integ[n] += Tln1pluseminepsinfty[n] + * (1.0 - (atan((Set.epsilon[n].lambdamax - Set.epsilon[0].lambda[Set.epsilon[0].Npts/2])/(0.5 * n * c_int)) + + atan((Set.epsilon[n].lambdamax + Set.epsilon[0].lambda[Set.epsilon[0].Npts/2])/(0.5 * n * c_int)))/PI); + // Prediction based on value_infty: + integ_target[n] = -kBT * log(1.0 - pow(sinh(Omega/kBT)/sinh((n + 1.0) * Omega/kBT), 2.0)); + integ_target[n] = -kBT * log(1.0 - pow((1.0 - exp(-2.0*Omega/kBT))/(1.0 - exp(-2.0*(n + 1.0)*Omega/kBT)), 2.0) * exp(-2.0 * n * Omega/kBT)); + cout << n << " " << Set.epsilon[n].value[0] << " " << Set.epsilon[n].value_infty << " " << " " << Set.epsilon[n].value[Set.epsilon[n].Npts/2] + << " " << Tln1plusemineps[n][0] << " " << Tln1plusemineps[n][Set.epsilon[n].Npts/2] << " " << Tln1pluseminepsinfty[n] + << " " << a_n_Tln_conv_0_integ[n] << " " << integ_target[n] << "\t"; + } + cout << endl; + //JSCerror("Stop..."); + */ + // Reconstruct the epsilon[0] function: + for (int i = 0; i < Set.epsilon[0].Npts; ++i) { + Set.epsilon[0].value[i] = pow(Set.epsilon[0].lambda[i], 2.0) - mu - Omega; + + // Add the convolutions: + for (int n = 0; n < Set.ntypes; ++n) + //if (n <= 1) // REMOVE + Set.epsilon[0].value[i] -= a_n_Tln_conv[n][i]; + + // Add the asymptotic parts of convolutions: + for (int n = 1; n < Set.ntypes; ++n) + //if (n <= 1) // REMOVE + Set.epsilon[0].value[i] -= Tln1pluseminepsinfty[n] + * (1.0 - (atan((Set.epsilon[n].lambdamax - Set.epsilon[0].lambda[i])/(0.5 * n * c_int)) + + atan((Set.epsilon[n].lambdamax + Set.epsilon[0].lambda[i])/(0.5 * n * c_int)))/PI); + + // Add the leftover summation for species n > nmax, assuming epsilon_n = epsilon_n^\infty in those cases: + Set.epsilon[0].value[i] -= Smaxsum; + + //cout << "i " << i << "\tlambda " << Set.epsilon[0].lambda[i] << "\te[0][i] " << Set.epsilon[0].value[i] << "\tprev " << Set.epsilon[0].prev_value[i] + // << "\tlambda^2 " << pow(Set.epsilon[0].lambda[i], 2.0) << "\ta_n_Tln_conv[0] " << a_n_Tln_conv[0][i] << "\ta_n_Tln_conv[1] " << a_n_Tln_conv[1][i] + // << endl; + //cout << a_n_dlambda[i][1] << endl << endl; + + // Include some damping: + Set.epsilon[0].value[i] = 0.1 * Set.epsilon[0].prev_value[i] + 0.9 * Set.epsilon[0].value[i]; + // No need to force boundaries here, epsilon[0] is inherently stable. + } + // epsilon[0] is now fully iterated. + + //cout << "Here 1" << endl; + + // Now do the remaining epsilons: + + for (int n = 1; n < Set.ntypes; ++n) { + + Vect_DP f_Tln_conv_min (0.0, Set.epsilon[n].Npts); // 'down' convolution + Vect_DP f_Tln_conv_plus (0.0, Set.epsilon[n].Npts); // 'up' convolution + + // For n = ntypes, need: + //DP Tln1pluseepsntypesinfty = 2.0 * Omega * Set.ntypes + //+ 2.0 * kBT * + //((Set.ntypes * Omega/kBT < 24.0 ? log(1.0 - exp(-2.0 * (Set.ntypes + 1) * Omega/kBT)) : - exp(-2.0 * (Set.ntypes + 1) * Omega/kBT)) + // - (2.0 * Omega/kBT < 24.0 ? log(1.0 - exp(-2.0 * Omega/kBT)) : - exp(-2.0 * Omega/kBT))); + + Vect_DP fmin(0.0, Set.epsilon[n-1].Npts); + Vect_DP fplus(0.0, Set.epsilon[JSC::min(n+1, Set.ntypes - 1)].Npts); + + for (int i = 0; i < Set.epsilon[n].Npts; ++i) { + f_Tln_conv_min[i] = 0.0; + f_Tln_conv_plus[i] = 0.0; + + // 'down' convolutions + if (n == 1) + for (int j = 0; j < Set.epsilon[0].Npts; ++j) + f_Tln_conv_min[i] += Tln1plusemineps[0][j] * fmin_dlambda[n][i][j]; + + else for (int j = 0; j < Set.epsilon[n - 1].Npts; ++j) + f_Tln_conv_min[i] += (Set.epsilon[n-1].prev_value[j] - Set.epsilon[n-1].value_infty + Tln1plusemineps[n-1][j] - Tln1pluseminepsinfty[n-1]) + * fmin_dlambda[n][i][j]; + + // 'up' convolutions + if (n < Set.ntypes - 1) + for (int j = 0; j < Set.epsilon[n+1].Npts; ++j) + f_Tln_conv_plus[i] += (Set.epsilon[n+1].prev_value[j] - Set.epsilon[n+1].value_infty + Tln1plusemineps[n+1][j] - Tln1pluseminepsinfty[n+1]) + * fplus_dlambda[n][i][j]; + + else f_Tln_conv_plus[i] = 0.0; + + // Do some damping: + Set.epsilon[n].value[i] = 0.1 * Set.epsilon[n].prev_value[i] + + 0.9 * (Set.epsilon[n].value_infty + f_Tln_conv_min[i] + f_Tln_conv_plus[i]); + // Force boundary values to asymptotes: force boundary 10 points on each side + if (i < 10) + Set.epsilon[n].value[i] = (1.0 - 0.1 * i) * Set.epsilon[n].value_infty + 0.1 * i * Set.epsilon[n].value[i]; + if (i > Set.epsilon[n].Npts - 11) + Set.epsilon[n].value[i] = (1.0 - 0.1 * (Set.epsilon[n].Npts-1 - i)) * Set.epsilon[n].value_infty + 0.1 * (Set.epsilon[n].Npts-1 - i) * Set.epsilon[n].value[i]; + /* + if (i == 0 && n < 2) { + cout << "epsilon[" << n << "][0]: " << Set.epsilon[n].value[i] << "\t" << Set.epsilon[n].prev_value[i] << "\t" + << Set.epsilon[n].value_infty << "\t" << f_Tln_conv_min[i] << "\t" << f_Tln_conv_plus[i] + //<< "\tepsilon[" << n << "][1]: " << Set.epsilon[n].value[1] << "\t" << Set.epsilon[n].prev_value[1] << "\t" + // << Set.epsilon[n].value_infty << "\t" << f_Tln_conv_min[1] << "\t" << f_Tln_conv_plus[1] + << "\tepsilon[" << n << "][10]: " << Set.epsilon[n].value[10] << "\t" << Set.epsilon[n].prev_value[10] << "\t" + << Set.epsilon[n].value_infty << "\t" << f_Tln_conv_min[10] << "\t" << f_Tln_conv_plus[10] + << endl; + } + */ + /* + if (n == 1) cout << setprecision(8) << Set.epsilon[n].lambda[i] << "\t" << Set.epsilon[n].prev_value[i] << "\t" << Set.epsilon[n].value[i] << "\t" + << setprecision(16) << f_Tln_conv_min[i] << "\t" << f_Tln_conv_plus[i] << "\t" + << fplus_dlambda[n][i][Set.epsilon[n+1].Npts - 1] + << endl; + */ + /* + if (fabs(1.0 - Set.epsilon[n].value[i]/Set.epsilon[n].prev_value[i]) > 0.1) { + cout << n << "\t" << i << "\t" << setprecision(8) << Set.epsilon[n].lambda[i] << "\t" << Set.epsilon[n].prev_value[i] << "\t" << Set.epsilon[n].value[i] << "\t" + << Set.epsilon[n].value_infty << "\t" << setprecision(16) << f_Tln_conv_min[i] << "\t" << f_Tln_conv_plus[i] + << endl; + cout << Set.epsilon[n+1].prev_value << endl << endl << Tln1plusemineps[n+1] << endl << endl; + cout << "dlambda: " << Set.epsilon[n+1].dlambda << endl << endl; + for (int j = 0; j < Set.epsilon[n+1].Npts; ++j) + cout << fplus_dlambda[n][i][j]/Set.epsilon[n+1].dlambda[j] << " "; + cout << endl << endl; + for (int j = 0; j < Set.epsilon[n+1].Npts; ++j) + cout << Set.epsilon[n+1].prev_value[j] - Set.epsilon[n+1].value_infty + Tln1plusemineps[n+1][j] - Tln1pluseminepsinfty[n+1] << " "; + cout << endl << endl; + } + */ + + } // for (int i = 0... + + } // for (int n = 1... + + // All functions have now been iterated. + + // Now calculate diff: + + DP eps0i = 0.0; + DP eps1i = 0.0; + + Set.diff = 0.0; + + for (int n = 0; n < Set.ntypes; ++n) { + Set.epsilon[n].diff = 0.0; + //sum_N += Set.epsilon[n].Npts; + for (int i = 0; i < Set.epsilon[n].Npts; ++i) { + //Set.epsilon[n].diff += pow((Set.epsilon[n].value[i] - Set.epsilon[n].prev_value[i]) + // /JSC::max(1.0, fabs(Set.epsilon[n].value[i] + Set.epsilon[n].prev_value[i])), 2.0); + //Set.epsilon[n].diff += fabs((Set.epsilon[n].value[i] - Set.epsilon[n].prev_value[i]) + // /JSC::max(1.0, fabs(Set.epsilon[n].value[i] + Set.epsilon[n].prev_value[i]))); + //Set.epsilon[n].diff += fabs(kBT * log((1.0 + exp(-fabs(Set.epsilon[n].value[i])/kBT))/(1.0 + exp(-fabs(Set.epsilon[n].prev_value[i])/kBT)))); + /* + // This one was used in working version before delta f measure: + Set.epsilon[n].diff += Set.epsilon[n].dlambda[i] * + fabs(Set.epsilon[n].value[i] > 0.0 ? kBT * log((1.0 + exp(-Set.epsilon[n].value[i]/kBT))/(1.0 + exp(-Set.epsilon[n].prev_value[i]/kBT))) + : (-Set.epsilon[n].value[i] + kBT * log(1.0 + exp(Set.epsilon[n].value[i]/kBT))) + - (-Set.epsilon[n].prev_value[i] + kBT * log(1.0 + exp(Set.epsilon[n].prev_value[i]/kBT)))); + */ + // Measure based on delta f/delta epsilon: + if (n == 0) + Set.epsilon[n].diff += Set.epsilon[n].dlambda[i] * + //(Set.epsilon[0].value[i] > 0.0 ? exp(-Set.epsilon[0].value[i]/kBT)/(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) + // : 1.0/(1.0 + exp(Set.epsilon[0].value[i]/kBT))) + //pow(Set.epsilon[0].value[i] > 0.0 ? + // exp(-Set.epsilon[0].value[i]/kBT)/(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) : 1.0/(1.0 + exp(Set.epsilon[0].value[i]/kBT)), 2.0) + //* fabs(Set.epsilon[n].value[i] - Set.epsilon[n].prev_value[i]); + (Set.epsilon[0].value[i] > 0.0 ? + exp(-Set.epsilon[0].value[i]/kBT)/(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) : 1.0/(1.0 + exp(Set.epsilon[0].value[i]/kBT))) + * fabs(Set.epsilon[n].value[i] - Set.epsilon[n].prev_value[i]); + else { + eps0i = Set.epsilon[0].Return_Value(Set.epsilon[n].lambda[i]); + eps1i = Set.epsilon[1].Return_Value(Set.epsilon[n].lambda[i]); + + Set.epsilon[n].diff += Set.epsilon[n].dlambda[i] * + // Logic: simple 1/2 cascade + (eps0i > 0.0 ? exp(-eps0i/kBT)/(1.0 + exp(-eps0i/kBT)) : 1.0/(1.0 + exp(eps0i/kBT))) + * pow(0.5, n) //* (exp(-eps1i/kBT)/(1.0 + exp(-eps1i/kBT))) + * fabs(Set.epsilon[n].value[i] - Set.epsilon[n].prev_value[i]); + } + //if (n == 0) cout << i << "\t" << Set.epsilon[n].value[i] << "\t" << Set.epsilon[n].prev_value[i] + // << "\t" << Set.epsilon[n].value[i] - Set.epsilon[n].prev_value[i] << "\t"; + //if (n == 0 && i == Set.epsilon[n].Npts - 1) cout << endl; + /* + fabs(kBT * ((Set.epsilon[n].value[i] < 24.0 * kBT ? log(1.0 + exp(-Set.epsilon[n].value[i]/kBT)) + : exp(-fabs(Set.epsilon[n].value[i])/kBT)) + - (fabs(Set.epsilon[n].prev_value[i]) < 24.0 * kBT ? log(1.0 + exp(-fabs(Set.epsilon[n].prev_value[i])/kBT)) + : exp(-fabs(Set.epsilon[n].prev_value[i])/kBT)))); + */ + } + //Set.epsilon[n].diff /= Set.epsilon[n].Npts; + Set.diff += Set.epsilon[n].diff; + //cout << n << " " << Set.epsilon[n].diff << "\t"; + } + //Set.diff /= Set.ntypes; + //cout << endl; + + return; +} + +void Iterate_and_Extrapolate_2CBG_TBAE (Root_Density_Set& Set, Vect& IterSet, Vect >& a_n_dlambda, + Vect >& fmin_dlambda, Vect >& fplus_dlambda, DP c_int, DP mu, DP Omega, DP kBT) +{ + // + + int nfit = IterSet.size(); + + for (int ifit = 0; ifit < nfit; ++ifit) { + Iterate_2CBG_TBAE (Set, a_n_dlambda, fmin_dlambda, fplus_dlambda, c_int, mu, Omega, kBT); + IterSet[ifit] = Set; + } + + // Now extrapolate each value to infinite nr of iterations: + Vect_DP density(nfit); + Vect_DP oneoverP(nfit); + DP deltalambda = 0.0; + for (int ifit = 0; ifit < nfit; ++ifit) oneoverP[ifit] = 1.0/(1.0 + ifit*ifit); + for (int n = 0; n < Set.ntypes; ++n) + for (int i = 0; i < Set.epsilon[n].Npts; ++i) { + for (int ifit = 0; ifit < nfit; ++ifit) density[ifit] = IterSet[ifit].epsilon[n].value[i]; + polint (oneoverP, density, 0.0, Set.epsilon[n].value[i], deltalambda); + } + + // Now iterate a few times to stabilize: + for (int iint = 0; iint < 2; ++iint) Iterate_2CBG_TBAE(Set, a_n_dlambda, fmin_dlambda, fplus_dlambda, c_int, mu, Omega, kBT); + + return; +} + +DP Refine_2CBG_Set (Root_Density_Set& Set, DP c_int, DP mu, DP Omega, DP kBT, DP refine_fraction) +{ + // This function replaces Set by a new set with more points, where + // Tln(...) needs to be evaluated more precisely. + + // The return value is the max of delta_tni found. + + // First, calculate the needed Tln... + Vect Tln1plusemineps(Set.ntypes); + Vect_DP Tln1pluseminepsinfty(Set.ntypes); + + for (int n = 0; n < Set.ntypes; ++n) { + + Tln1plusemineps[n] = Vect_DP (0.0, Set.epsilon[n].Npts); + + for (int i = 0; i < Set.epsilon[n].Npts; ++i) { + Tln1plusemineps[n][i] = Set.epsilon[n].value[i] > 0.0 ? + kBT * (Set.epsilon[n].value[i] < 24.0 * kBT ? log(1.0 + exp(-Set.epsilon[n].value[i]/kBT)) : exp(-Set.epsilon[n].value[i]/kBT)) + : -Set.epsilon[n].value[i] + kBT * log (1.0 + exp(Set.epsilon[n].value[i]/kBT)); + } + Tln1pluseminepsinfty[n] = kBT * (Set.epsilon[n].value_infty < 24.0 * kBT ? + log(1.0 + exp(-Set.epsilon[n].value_infty/kBT)) : exp(-Set.epsilon[n].value_infty/kBT)); + } + + // Now find the achieved delta_tni + DP max_delta_tni_dlambda = 0.0; + DP max_delta_tni_dlambda_toplevel = 0.0; + DP sum_delta_tni_dlambda = 0.0; + //DP tni = 0.0; + + Vect tni(Set.ntypes); + Vect tni_ex(Set.ntypes); + + //Vect_DP delta_tni_dlambda(0.0, Set.Npts_total); + //int delta_tni_counter = 0; + + DP measure_factor = 0.0; + DP eps0i = 0.0; + DP eps1i = 0.0; + + for (int n = 0; n < Set.ntypes; ++n) { + + tni[n] = Vect_DP (0.0, Set.epsilon[n].Npts); + tni_ex[n] = Vect_DP (0.0, Set.epsilon[n].Npts); // extrapolation from adjacent points, to compare to obtained value + + for (int i = 1; i < Set.epsilon[n].Npts - 1; ++i) { + //tni[n][i] = Tln1plusemineps[n][i] - Tln1pluseminepsinfty[n]; + //tni_ex[n][i] = ((Tln1plusemineps[n][i-1] - Tln1pluseminepsinfty[n]) * (Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i]) + // + (Tln1plusemineps[n][i+1] - Tln1pluseminepsinfty[n]) * (Set.epsilon[n].lambda[i] - Set.epsilon[n].lambda[i-1])) + ///(Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i-1]); + if (n == 0) { + /* + tni[n][i] = Set.epsilon[n].value[i] - pow(Set.epsilon[n].lambda[i], 2.0) + mu + Omega + Tln1plusemineps[n][i] - Tln1pluseminepsinfty[n]; + tni_ex[n][i] = ((Set.epsilon[n].value[i-1] - pow(Set.epsilon[n].lambda[i-1], 2.0) + mu + Omega + + Tln1plusemineps[n][i-1] - Tln1pluseminepsinfty[n]) * (Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i]) + + (Set.epsilon[n].value[i+1] - pow(Set.epsilon[n].lambda[i+1], 2.0) + mu + Omega + + Tln1plusemineps[n][i+1] - Tln1pluseminepsinfty[n]) * (Set.epsilon[n].lambda[i] - Set.epsilon[n].lambda[i-1])) + /(Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i-1]); + */ + /* + // Working version: + tni[n][i] = Tln1plusemineps[n][i]; + tni_ex[n][i] = (Tln1plusemineps[n][i-1] * (Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i]) + + Tln1plusemineps[n][i+1] * (Set.epsilon[n].lambda[i] - Set.epsilon[n].lambda[i-1])) + /(Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i-1]); + */ + // Measure based on delta f/delta epsilon: + //measure_factor = Set.epsilon[0].value[i] > 0.0 ? + //exp(-Set.epsilon[0].value[i]/kBT)/(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) : 1.0/(1.0 + exp(Set.epsilon[0].value[i]/kBT)); + //measure_factor = pow(Set.epsilon[0].value[i] > 0.0 ? exp(-Set.epsilon[0].value[i]/kBT)/(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) + // : 1.0/(1.0 + exp(Set.epsilon[0].value[i]/kBT)), 2.0); + measure_factor = (Set.epsilon[0].value[i] > 0.0 ? exp(-Set.epsilon[0].value[i]/kBT)/(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) + : 1.0/(1.0 + exp(Set.epsilon[0].value[i]/kBT))); + } + else { + /* + // This is the more natural choice, since delta(epsilon[n]) gets transferred more or less linearly to delta(epsilon[0]): + tni[n][i] = Set.epsilon[n].value[i] - Set.epsilon[n].value_infty + Tln1plusemineps[n][i] - Tln1pluseminepsinfty[n]; + tni_ex[n][i] = ((Set.epsilon[n].value[i-1] - Set.epsilon[n].value_infty + Tln1plusemineps[n][i-1] - Tln1pluseminepsinfty[n]) + * (Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i]) + + (Set.epsilon[n].value[i+1] - Set.epsilon[n].value_infty + Tln1plusemineps[n][i+1] - Tln1pluseminepsinfty[n]) + * (Set.epsilon[n].lambda[i] - Set.epsilon[n].lambda[i-1])) + /(Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i-1]); + */ + /* + tni[n][i] = Tln1plusemineps[n][i] - Tln1pluseminepsinfty[n]; + tni_ex[n][i] = ((Tln1plusemineps[n][i-1] - Tln1pluseminepsinfty[n]) * (Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i]) + + (Tln1plusemineps[n][i+1] - Tln1pluseminepsinfty[n]) * (Set.epsilon[n].lambda[i] - Set.epsilon[n].lambda[i-1])) + /(Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i-1]); + */ + // Measure based on delta f/delta epsilon: + eps0i = Set.epsilon[0].Return_Value(Set.epsilon[n].lambda[i]); + eps1i = Set.epsilon[1].Return_Value(Set.epsilon[n].lambda[i]); + + measure_factor = (eps0i > 0.0 ? exp(-eps0i/kBT)/(1.0 + exp(-eps0i/kBT)) : 1.0/(1.0 + exp(eps0i/kBT))) + // Logic: simple 1/2 per level cascade down + //* (exp(-eps1i/kBT)/(1.0 + exp(-eps1i/kBT))) + * pow(0.5, n); + + } + + tni[n][i] = measure_factor * Set.epsilon[n].value[i]; + tni_ex[n][i] = measure_factor * (Set.epsilon[n].value[i-1] * (Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i]) + + Set.epsilon[n].value[i+1] * (Set.epsilon[n].lambda[i] - Set.epsilon[n].lambda[i-1])) + /(Set.epsilon[n].lambda[i+1] - Set.epsilon[n].lambda[i-1]); + + max_delta_tni_dlambda = JSC::max(max_delta_tni_dlambda, fabs(tni[n][i] - tni_ex[n][i]) * Set.epsilon[n].dlambda[i]); + if (n == Set.ntypes - 1) + max_delta_tni_dlambda_toplevel = JSC::max(max_delta_tni_dlambda_toplevel, fabs(tni[n][i] - tni_ex[n][i]) * Set.epsilon[n].dlambda[i]); + sum_delta_tni_dlambda += fabs(tni[n][i] - tni_ex[n][i]) * Set.epsilon[n].dlambda[i]; + } + } + + //cout << "Before sort: " << endl << delta_tni_dlambda << endl; + //delta_tni_dlambda.QuickSort(); + //cout << "After sort: " << endl << delta_tni_dlambda << endl; + //max_delta_tni_dlambda = delta_tni_dlambda[int(delta_tni_dlambda.size() * (1.0 - refine_fraction))]; + //cout << "max_delta_tni_dlambda = " << max_delta_tni_dlambda << endl;//"\tindex = " << int(delta_tni_dlambda.size() * (1.0 - refine_fraction)) << endl; + + // We now determine the locations where we need to add points + Vect > need_new_point_around(Set.ntypes); + Vect need_to_extend_limit(false, Set.ntypes); + + for (int n = 0; n < Set.ntypes; ++n) { + + need_new_point_around[n] = Vect (false, Set.epsilon[n].Npts); + + for (int i = 1; i < Set.epsilon[n].Npts - 1; ++i) { + if (fabs(tni[n][i] - tni_ex[n][i]) * Set.epsilon[n].dlambda[i] > (1.0 - refine_fraction) * max_delta_tni_dlambda) { + need_new_point_around[n][i] = true; + // Do also the symmetric ones... Require need...[n][i] = need...[n][Npts - 1 - i] + need_new_point_around[n][Set.epsilon[n].Npts - 1 - i] = true; + } + } + + // Check boundary values; if too different from value_infty, extend limits + if (n == 0) { + //if (fabs(Tln1plusemineps[0][0]/Tln1pluseminepsinfty[n] - 1.0) * Set.epsilon[0].dlambda[0] > (1.0 - refine_fraction) * max_delta_tni_dlambda) + //if (fabs(Tln1plusemineps[0][0]/Tln1pluseminepsinfty[n] - 1.0) > (1.0 - refine_fraction) * max_delta_tni_dlambda) + //if (Tln1plusemineps[0][0] > (1.0 - refine_fraction) * max_delta_tni_dlambda) + // Used in working version: + //if (10000.0 * Tln1plusemineps[0][0] > max_delta_tni_dlambda/Set.epsilon[0].Npts) + //need_to_extend_limit[0] = true; + // Measure based on delta f/delta epsilon: + if (exp(-Set.epsilon[0].value[0]/kBT) > 0.001 * max_delta_tni_dlambda) + need_to_extend_limit[0] = true; + } + else + //if (fabs(Set.epsilon[n].value[0]/Set.epsilon[n].value_infty - 1.0) * Set.epsilon[n].dlambda[0] > (1.0 - refine_fraction) * max_delta_tni_dlambda) + //if (fabs(Set.epsilon[n].value[0]/Set.epsilon[n].value_infty - 1.0) > (1.0 - refine_fraction) * max_delta_tni_dlambda) + // Used in working version: + //if (10000.0 * fabs(Tln1plusemineps[n][0] - Tln1pluseminepsinfty[n]) > max_delta_tni_dlambda/Set.epsilon[n].Npts) + // Measure deviation from asymptote for 10th element, since we smoothly put the i < 10 ones to the asymptote when damping: + if (fabs(Set.epsilon[n].value[10] - Set.epsilon[n].value_infty) * Set.epsilon[0].dlambda[10] > max_delta_tni_dlambda) + need_to_extend_limit[n] = true; + } + + // Check if we need to add a level + bool need_new_epsilon_n_function = false; + + // We add new levels if the integral a_n * Tln1plusemineps at the highest level differs too much from + // the asymptotic value. Since such integrals appear for each point of the epsilon[0] function, these + // errors should be compared to the individual delta_tni factors. + DP a_2_Tln_conv_0_integ = 0.0; + DP oneoverpi = 1.0/PI; + DP twoovernc = 2.0/((Set.ntypes - 1) * c_int); + int i0 = Set.epsilon[0].Npts/2; + for (int j = 0; j < Set.epsilon[Set.ntypes - 1].Npts; ++j) + a_2_Tln_conv_0_integ += (Tln1plusemineps[Set.ntypes - 1][j] - Tln1pluseminepsinfty[Set.ntypes - 1]) + * oneoverpi * (atan(twoovernc * (Set.epsilon[0].lambda[i0] - Set.epsilon[Set.ntypes - 1].lambda[j] + 0.5 * Set.epsilon[Set.ntypes - 1].dlambda[j])) + - atan(twoovernc * (Set.epsilon[0].lambda[i0] - Set.epsilon[Set.ntypes - 1].lambda[j] - 0.5 * Set.epsilon[Set.ntypes - 1].dlambda[j]))); + // Add asymptotic parts: not necessary, identically 0 + //a_2_Tln_conv_0_integ += Tln1pluseminepsinfty[Set.ntypes - 1] + //* (1.0 - (atan((Set.epsilon[Set.ntypes - 1].lambdamax - Set.epsilon[0].lambda[i0])/(0.5 * (Set.ntypes - 1) * c_int)) + // + atan((Set.epsilon[Set.ntypes - 1].lambdamax + Set.epsilon[0].lambda[i0])/(0.5 * (Set.ntypes - 1) * c_int)))/PI); + + // Compare to prediction for this integral based on value_infty, which is simply 0. + // Count this difference Set.ntypes times over, since it cascades down all levels + if (fabs(a_2_Tln_conv_0_integ) * Set.ntypes > max_delta_tni_dlambda) need_new_epsilon_n_function = true; + + //cout << "Toplevel check: fabs(integ)* ntypes = " << fabs(a_2_Tln_conv_0_integ) * Set.ntypes + // << "\tmax tni = " << max_delta_tni_dlambda << "\tbool: " << (fabs(a_2_Tln_conv_0_integ) * Set.ntypes > max_delta_tni_dlambda) << endl; + + // Additionally, if the highest level needs updating, we automatically add new functions: + for (int i = 0; i < Set.epsilon[Set.ntypes - 1].Npts; ++i) + if (need_new_point_around[Set.ntypes - 1][i] || need_to_extend_limit[Set.ntypes - 1]) need_new_epsilon_n_function = true; + + // Finally, we also add functions if epsilon[n] itself is too different from epsilon[n](infty), + // based on delta f/delta epsilon_n measure: + //if (fabs(Set.epsilon[Set.ntypes - 1].value[Set.epsilon[Set.ntypes - 1].Npts/2] - Set.epsilon[Set.ntypes - 1].value_infty) + // Next expression: replaced by following one + //* exp(-(Set.epsilon[0].value[Set.epsilon[0].Npts/2] + Set.epsilon[Set.ntypes - 1].value[Set.epsilon[Set.ntypes - 1].Npts/2])/kBT) + ///(1.0 + exp(-Set.epsilon[Set.ntypes - 1].value[Set.epsilon[Set.ntypes - 1].Npts/2]/kBT)) + //* exp(-Set.epsilon[Set.ntypes - 1].value[Set.epsilon[Set.ntypes - 1].Npts/2]/kBT) + // /((1.0 + exp(Set.epsilon[0].value[Set.epsilon[0].Npts/2]/kBT)) * (1.0 + exp(-Set.epsilon[Set.ntypes - 1].value[Set.epsilon[Set.ntypes - 1].Npts/2]/kBT))) + // Logic: simple 1/2 factor cascade + //* pow(0.5, Set.ntypes) + //* exp(-Set.epsilon[1].value[Set.epsilon[1].Npts/2]/kBT) + ///((1.0 + exp(Set.epsilon[0].value[Set.epsilon[0].Npts/2]/kBT)) * (1.0 + exp(-Set.epsilon[1].value[Set.epsilon[1].Npts/2]/kBT))) + // > max_delta_tni_dlambda) need_new_epsilon_n_function = true; + /* + cout << "New level ? " << fabs(Set.epsilon[Set.ntypes - 1].value[Set.epsilon[Set.ntypes - 1].Npts/2] - Set.epsilon[Set.ntypes - 1].value_infty) + << "\t" << pow(0.5, Set.ntypes) << "\t" << fabs(Set.epsilon[Set.ntypes - 1].value[Set.epsilon[Set.ntypes - 1].Npts/2] - Set.epsilon[Set.ntypes - 1].value_infty) + * pow(0.5, Set.ntypes) << "\t" << max_delta_tni_dlambda + << "\t" << (fabs(Set.epsilon[Set.ntypes - 1].value[Set.epsilon[Set.ntypes - 1].Npts/2] - Set.epsilon[Set.ntypes - 1].value_infty) + * pow(0.5, Set.ntypes) > max_delta_tni_dlambda) + << "\t" << need_new_epsilon_n_function + << "\t" + //<< endl; + ; + */ + //if (max_delta_tni_dlambda_toplevel > pow(0.5, Set.ntypes) * max_delta_tni_dlambda && Set.ntypes < 100) need_new_epsilon_n_function = true; + //cout << "New level above top level ? " << max_delta_tni_dlambda_toplevel << "\t" << max_delta_tni_dlambda << "\t" + // << (max_delta_tni_dlambda_toplevel > 0.01 * max_delta_tni_dlambda) << endl; + + // Now insert the new points between existing points: + Set.Insert_new_points (need_new_point_around); + + // Now extend the integration limits if needed: + Set.Extend_limits (need_to_extend_limit); + + // If we add a level, we do it here: + if (need_new_epsilon_n_function) { + // Insert more than one function per cycle... + Set.Insert_new_function(Asymptotic_2CBG_epsilon(Set.ntypes, Omega, kBT)); + Set.Insert_new_function(Asymptotic_2CBG_epsilon(Set.ntypes, Omega, kBT)); // CAREFUL !! ntypes is already updated + Set.Insert_new_function(Asymptotic_2CBG_epsilon(Set.ntypes, Omega, kBT)); // CAREFUL !! ntypes is already updated + Set.Insert_new_function(Asymptotic_2CBG_epsilon(Set.ntypes, Omega, kBT)); // CAREFUL !! ntypes is already updated; + Set.Insert_new_function(Asymptotic_2CBG_epsilon(Set.ntypes, Omega, kBT)); // CAREFUL !! ntypes is already updated + } + + //return(max_delta_tni_dlambda); + return(sum_delta_tni_dlambda); +} + +DP Calculate_Gibbs_Free_Energy (const Root_Density_Set& Set, DP c_int, DP mu, DP Omega, DP kBT) +{ + // Computes the Gibbs free energy, assuming that epsilon[0] is symmetric. + + // WORKING VERSION + DP sum_f = 0.0; + Vect_DP f(0.0, Set.epsilon[0].Npts); + DP sum_f_check = 0.0; + Vect_DP fcheck(0.0, Set.epsilon[0].Npts); + DP sum_g_check = 0.0; + Vect_DP gcheck(0.0, Set.epsilon[0].Npts); + for (int i = 0; i < Set.epsilon[0].Npts; ++i) { + f[i] = (Set.epsilon[0].value[i] > 0.0 ? kBT* log(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) + : -Set.epsilon[0].value[i] + kBT * log(1.0 + exp(Set.epsilon[0].value[i]/kBT))); + sum_f += Set.epsilon[0].dlambda[i] * f[i]; + fcheck[i] = kBT * log(1.0 + exp(-(Set.epsilon[0].lambda[i] * Set.epsilon[0].lambda[i] - mu - Omega)/kBT)); + sum_f_check += Set.epsilon[0].dlambda[i] * fcheck[i]; + gcheck[i] = exp(-(Set.epsilon[0].lambda[i] * Set.epsilon[0].lambda[i])/kBT); + sum_g_check += Set.epsilon[0].dlambda[i] * gcheck[i]; + } + + // Now add alpha curvature terms: + for (int i = 1; i < Set.epsilon[0].Npts - 1; ++i) + sum_f += pow(Set.epsilon[0].dlambda[i], 3.0) * ((f[i+1] - f[i])/(Set.epsilon[0].lambda[i+1] - Set.epsilon[0].lambda[i]) + - (f[i] - f[i-1])/(Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[i-1])) + /(12.0 * (Set.epsilon[0].lambda[i+1] - Set.epsilon[0].lambda[i-1])); + + // Now add alpha curvature terms: + DP sum_gcorralphacheck = 0.0; + Vect_DP gcorr_alpha_check(0.0, Set.epsilon[0].Npts); + for (int i = 1; i < Set.epsilon[0].Npts - 1; ++i) { + //gcorr_alpha_check[i] = (pow(Set.epsilon[0].lambda[i+1] - Set.epsilon[0].lambda[i], 3.0) - pow(Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[i-1], 3.0)) + //* ((gcheck[i+1] - gcheck[i]) * (Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[i-1]) + // - (gcheck[i] - gcheck[i-1]) * (Set.epsilon[0].lambda[i+1] - Set.epsilon[0].lambda[i])) + ///(24.0 * (Set.epsilon[0].lambda[i+1] - Set.epsilon[0].lambda[i-1]) * (Set.epsilon[0].lambda[i+1] - Set.epsilon[0].lambda[i]) + //* (Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[i-1])); + gcorr_alpha_check[i] = (1.0/12.0) * pow(Set.epsilon[0].dlambda[i], 3.0) + * ((gcheck[i+1] - gcheck[i]) * (Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[i-1]) + - (gcheck[i] - gcheck[i-1]) * (Set.epsilon[0].lambda[i+1] - Set.epsilon[0].lambda[i])) + /((Set.epsilon[0].lambda[i+1] - Set.epsilon[0].lambda[i-1]) * (Set.epsilon[0].lambda[i+1] - Set.epsilon[0].lambda[i]) + * (Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[i-1])); + sum_gcorralphacheck += gcorr_alpha_check[i]; + } + //sum_f += sum_fcorr + sum_fcorralpha; + //cout << sum_f << "\t" << sum_fcorr << "\t" << sum_fcorralpha << "\tRelative corr: " << sum_fcorr/sum_f << "\t" << sum_fcorralpha/sum_f << endl; + + /* + // Original version: + DP sum_f = 0.0; + for (int i = 0; i < Set.epsilon[0].Npts; ++i) { + sum_f += Set.epsilon[0].dlambda[i] * (Set.epsilon[0].value[i] > 0.0 ? kBT* log(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) + : -Set.epsilon[0].value[i] + kBT * log(1.0 + exp(Set.epsilon[0].value[i]/kBT))); + //cout << Set.epsilon[0].lambda[i] << "\t" << Set.epsilon[0].value[i] << "\t" << (Set.epsilon[0].value[i] > 0.0 ? kBT * log(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) + //: -Set.epsilon[0].value[i] + kBT * log(1.0 + exp(Set.epsilon[0].value[i]/kBT))) << endl; + } + */ + + // Testing: + int Npts_test = Set.epsilon[0].Npts; + DP lambdamax_test = Set.epsilon[0].lambdamax; + DP testsum = 0.0; + DP testgauss = 0.0; + DP lambda; + DP dlambda; + for (int i = 0; i < Npts_test; ++i) { + lambda = lambdamax_test * (-Npts_test + 1.0 + 2*i)/Npts_test; + dlambda = lambdamax_test * 2.0/Npts_test; + testsum += kBT * log(1.0 + exp(-(lambda*lambda - mu - Omega)/kBT)) * dlambda; + testgauss += exp(-lambda*lambda/kBT) * dlambda; + } + + //cout << "Test1: " << Npts_test << "\t" << lambdamax_test << "\t" << -testsum/twoPI << "\t" << testgauss/sqrt(PI * kBT) << endl; + /* + cout << setprecision(16) << "Npts = " << Set.epsilon[0].Npts << "\tlambdamax = " << Set.epsilon[0].lambdamax + << "\tf = " << -sum_f/twoPI << "\tGaussian: " << sum_g_check/sqrt(PI * kBT) << "\ttestGauss = " << testgauss/sqrt(PI * kBT) << endl; + cout << setprecision(16) << Set.epsilon[0].dlambda.sum() << "\t" << Set.epsilon[0].lambdamax * 2.0 << endl; + cout << gcheck << endl << endl; + */ + /* + // Output a test file: + ofstream outfile_test; + outfile_test.open("Test_Gauss.dat"); + outfile_test.precision(16); + DP gauss1 = 0.0; + DP gauss1beta = 0.0; + DP gauss1alpha = 0.0; + DP gauss2 = 0.0; + for (int i = 0; i < Npts_test; ++i) { + lambda = lambdamax_test * (-Npts_test + 1.0 + 2*i)/Npts_test; + gauss1 += Set.epsilon[0].dlambda[i] * gcheck[i]; + //gauss1beta += Set.epsilon[0].dlambda[i] * gcheck[i] + gcorr_check[i]; + gauss1alpha += Set.epsilon[0].dlambda[i] * gcheck[i] + gcorr_alpha_check[i]; + gauss2 += dlambda * exp(-lambda*lambda/kBT); + outfile_test << Set.epsilon[0].lambda[i] << "\t" << Set.epsilon[0].dlambda[i] << "\t" << gcheck[i] << "\t" << gauss1 << "\t" << gauss1alpha + << "\t" << lambda << "\t" << exp(-lambda*lambda/kBT) << "\t" << gauss2 << endl; + } + outfile_test.close(); + char a; + //cout << "Done..." << endl; + //cin >> a; + */ + /* + Npts_test = Set.epsilon[0].Npts * 2; + lambdamax_test = Set.epsilon[0].lambdamax; + testsum = 0.0; + testgauss = 0.0; + + for (int i = 0; i < Npts_test; ++i) { + lambda = lambdamax_test * (-Npts_test + 1.0 + 2*i)/Npts_test; + dlambda = lambdamax_test * 2.0/Npts_test; + testsum += kBT * log(1.0 + exp(-(lambda*lambda - mu - Omega)/kBT)) * dlambda; + testgauss += exp(-lambda*lambda/kBT) * dlambda; + } + + //cout << "Test2: " << Npts_test << "\t" << lambdamax_test << "\t" << -testsum/twoPI << "\t" << testgauss/sqrt(PI * kBT) << endl; + + Npts_test = Set.epsilon[0].Npts * 2; + lambdamax_test = Set.epsilon[0].lambdamax; + testsum = 0.0; + testgauss = 0.0; + + for (int i = 0; i < Npts_test; ++i) { + lambda = lambdamax_test * (-Npts_test + 1.0 + 2*i)/Npts_test; + dlambda = lambdamax_test * 2.0/Npts_test; + testsum += kBT * log(1.0 + exp(-(lambda*lambda - mu - Omega)/kBT)) * dlambda; + testgauss += exp(-lambda*lambda/kBT) * dlambda; + } + + //cout << "Test3: " << Npts_test << "\t" << lambdamax_test << "\t" << -testsum/twoPI << "\t" << testgauss/sqrt(PI * kBT) << endl; + + Npts_test = Set.epsilon[0].Npts * 2; + lambdamax_test = Set.epsilon[0].lambdamax; + testsum = 0.0; + testgauss = 0.0; + + for (int i = 0; i < Npts_test; ++i) { + lambda = lambdamax_test * (-Npts_test + 1.0 + 2*i)/Npts_test; + dlambda = lambdamax_test * (-Npts_test + 1.0 + 2*i + 2.0)/Npts_test - lambdamax_test * (-Npts_test + 1.0 + 2*i)/Npts_test; + testsum += kBT * log(1.0 + exp(-(lambda*lambda - mu - Omega)/kBT)) * dlambda; + testgauss += exp(-lambda*lambda/kBT) * dlambda; + } + + //cout << "Test4: " << Npts_test << "\t" << lambdamax_test << "\t" << -testsum/twoPI << "\t" << testgauss/sqrt(PI * kBT) << endl; + + Npts_test = Set.epsilon[0].Npts * 4; + lambdamax_test = Set.epsilon[0].lambdamax * 1; + testsum = 0.0; + testgauss = 0.0; + + for (int i = 0; i < Npts_test; ++i) { + lambda = lambdamax_test * (-Npts_test + 1.0 + 2*i)/Npts_test; + dlambda = lambdamax_test * 2.0/Npts_test; + testsum += kBT * log(1.0 + exp(-(lambda*lambda - mu - Omega)/kBT)) * dlambda; + testgauss += exp(-lambda*lambda/kBT) * dlambda; + } + + //cout << "Test5: " << Npts_test << "\t" << lambdamax_test << "\t" << -testsum/twoPI << "\t" << testgauss/sqrt(PI * kBT) << endl; + */ + + return(-sum_f/twoPI); +} + +DP Calculate_dGibbs_dchempot (const Root_Density_Set& DSet, const Root_Density_Set& Set, DP c_int, DP mu, DP Omega, DP kBT) +{ + // This calculates the derivative of the Gibbs free energy with respect to either of the two chemical potential, + // given the fundametal set Set for eps and DSet for either deps_du or deps_dOmega. + + DP sum_f = 0.0; + Vect_DP f(0.0, Set.epsilon[0].Npts); + + for (int i = 0; i < Set.epsilon[0].Npts; ++i) { + f[i] = DSet.epsilon[0].value[i] * (Set.epsilon[0].value[i] > 0.0 ? exp(-Set.epsilon[0].value[i]/kBT)/(1.0 + exp(-Set.epsilon[0].value[i]/kBT)) + : 1.0/(1.0 + exp(Set.epsilon[0].value[i]/kBT))); + sum_f += DSet.epsilon[0].dlambda[i] * f[i]; + } + + // Now add alpha curvature terms: + for (int i = 1; i < DSet.epsilon[0].Npts - 1; ++i) + sum_f += pow(DSet.epsilon[0].dlambda[i], 3.0) * ((f[i+1] - f[i])/(DSet.epsilon[0].lambda[i+1] - DSet.epsilon[0].lambda[i]) + - (f[i] - f[i-1])/(DSet.epsilon[0].lambda[i] - DSet.epsilon[0].lambda[i-1])) + /(12.0 * (DSet.epsilon[0].lambda[i+1] - DSet.epsilon[0].lambda[i-1])); + + return(sum_f/twoPI); +} + +//void Build_a_n_dlambda (Vect >& a_n_dlambda, const Root_Density_Set& Set, DP c_int, DP mu, DP Omega, DP kBT) +Vect > Build_a_n_dlambda (const Root_Density_Set& Set, DP c_int, DP mu, DP Omega, DP kBT) +{ + DP oneoverpi = 1.0/PI; + DP oneoverc = 1.0/c_int; + //DP oneovertwoc = 0.5/c_int; + //DP oneovercsq = oneoverc * oneoverc; + DP twoovernc = 2.0/c_int; + //DP oneovernc; + //DP oneoverncsq; + + // First define some useful functions: + //Vect midlambdaleft(Set.ntypes); // 0.5 * (lambda[n][j-1] + lambda[n][j]) + + for (int n = 0; n < Set.ntypes; ++n) { + + //midlambdaleft[n] = Vect_DP (0.0, Set.epsilon[n].Npts); + + for (int i = 1; i < Set.epsilon[n].Npts; ++i) { + ////midlambdaleft[n][i] = 0.5 * (Set.epsilon[n].lambda[i-1] + Set.epsilon[n].lambda[i]); + //midlambdaleft[n][i] = Set.epsilon[n].lambda[i] - 0.5 * Set.epsilon[n].dlambda[i]); + } + } // for (int n = 0...) + + //Vect > a_n_dlambda(Set.epsilon[0].Npts); + + Vect > a_n_dlambda(Set.epsilon[0].Npts); + for (int i = 0; i < Set.epsilon[0].Npts; ++i) { + a_n_dlambda[i] = Vect(Set.ntypes); + for (int n = 0; n < Set.ntypes; ++n) { + a_n_dlambda[i][n] = Vect_DP(0.0, Set.epsilon[n].Npts); + } + } + + for (int i = 0; i < Set.epsilon[0].Npts; ++i) { + //a_n_dlambda[i] = Vect (Set.ntypes); + + // Do n == 0 separately: + //a_n_dlambda[i][0] = Vect_DP (0.0, Set.epsilon[0].Npts); + /* + a_n_dlambda[i][0][0] = Set.epsilon[0].dlambda[0] * oneoverpi * c_int/(pow(Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[0], 2.0) + c_int*c_int); + a_n_dlambda[i][0][Set.epsilon[0].Npts - 1] = Set.epsilon[0].dlambda[Set.epsilon[0].Npts - 1] * oneoverpi + * c_int/(pow(Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[Set.epsilon[0].Npts - 1], 2.0) + c_int*c_int); + for (int j = 1; j < Set.epsilon[0].Npts - 1; ++j) + //a_n_dlambda[i][0][j] = oneoverpi * (atan((Set.epsilon[0].lambda[i] - midlambdaleft[0][j]) * oneoverc) + // - atan((Set.epsilon[0].lambda[i] - midlambdaleft[0][j+1]) * oneoverc)); + a_n_dlambda[i][0][j] = oneoverpi * atan(oneovertwoc * (Set.epsilon[0].lambda[j+1] - Set.epsilon[0].lambda[j-1]) + /(1.0 + oneovercsq * (Set.epsilon[0].lambda[i] - midlambdaleft[0][j]) + * (Set.epsilon[0].lambda[i] - midlambdaleft[0][j+1]))); + */ + for (int j = 0; j < Set.epsilon[0].Npts; ++j) + // CANNOT USE THE NEXT FORM !! a_n_dlambda is then possibly negative. + //a_n_dlambda[i][0][j] = oneoverpi * atan(oneoverc * Set.epsilon[0].dlambda[j] + // /(1.0 + oneovercsq * (Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[j] - 0.5 * Set.epsilon[0].dlambda[j]) + // * (Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[j] + 0.5 * Set.epsilon[0].dlambda[j]))); + a_n_dlambda[i][0][j] = oneoverpi * (atan(oneoverc * (Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[j] + 0.5 * Set.epsilon[0].dlambda[j])) + - atan(oneoverc * (Set.epsilon[0].lambda[i] - Set.epsilon[0].lambda[j] - 0.5 * Set.epsilon[0].dlambda[j]))); + + //if (fabs(a_n_dlambda[i][0].sum() - 1.0) > 1.0e-12) cout << "Error in sum a_n for n = " << 0 << " i = " << i << "\tsum = " + // << a_n_dlambda[i][0].sum() << endl << endl << a_n_dlambda[i][0] << endl << endl; + + // Now do n > 0: + for (int n = 1; n < Set.ntypes; ++n) { + //a_n_dlambda[i][n] = Vect_DP (0.0, Set.epsilon[n].Npts); + + //a_n_dlambda[i][n][0] = Set.epsilon[n].dlambda[0] * 0.5 * n * oneoverpi * c_int + ///(pow(Set.epsilon[0].lambda[i] - Set.epsilon[n].lambda[0], 2.0) + 0.25 * n * n * c_int*c_int); + //a_n_dlambda[i][n][Set.epsilon[n].Npts - 1] = Set.epsilon[n].dlambda[Set.epsilon[n].Npts - 1] * 0.5 * n * oneoverpi * c_int + // /(pow(Set.epsilon[0].lambda[i] - Set.epsilon[n].lambda[Set.epsilon[n].Npts - 1], 2.0) + 0.25 * n * n * c_int*c_int); + twoovernc = 2.0/(n * c_int); + //oneovernc = 1.0/(n * c_int); + //oneoverncsq = oneovernc * oneovernc; + //for (int j = 1; j < Set.epsilon[n].Npts - 1; ++j) { + for (int j = 0; j < Set.epsilon[n].Npts; ++j) { + //a_n_dlambda[i][n][j] = oneoverpi * (atan(twoovernc * (Set.epsilon[0].lambda[i] - midlambdaleft[n][j])) + // - atan(twoovernc * (Set.epsilon[0].lambda[i] - midlambdaleft[n][j+1]))); + // CANNOT USE THE NEXT FORM !! a_n_dlambda is then possibly negative. + //a_n_dlambda[i][n][j] = oneoverpi * atan(oneovernc * (Set.epsilon[n].lambda[j+1] - Set.epsilon[n].lambda[j-1]) + // /(1.0 + oneoverncsq * (2.0 * Set.epsilon[0].lambda[i] - Set.epsilon[n].lambda[j-1] - Set.epsilon[n].lambda[j]) + // * (2.0 * Set.epsilon[0].lambda[i] - Set.epsilon[n].lambda[j+1] - Set.epsilon[n].lambda[j]))); + a_n_dlambda[i][n][j] = oneoverpi * (atan(twoovernc * (Set.epsilon[0].lambda[i] - Set.epsilon[n].lambda[j] + 0.5 * Set.epsilon[n].dlambda[j])) + - atan(twoovernc * (Set.epsilon[0].lambda[i] - Set.epsilon[n].lambda[j] - 0.5 * Set.epsilon[n].dlambda[j]))); + } + } // for (int n + } // for (int i + + return(a_n_dlambda); +} + +//void Build_fmin_dlambda (Vect >& fmin_dlambda, const Root_Density_Set& Set, DP c_int, DP mu, DP Omega, DP kBT) +Vect > Build_fmin_dlambda (const Root_Density_Set& Set, DP c_int, DP mu, DP Omega, DP kBT) +{ + DP oneoverpi = 1.0/PI; + //DP oneoverc = 1.0/c_int; + //DP twoovernc = 2.0/c_int; + DP pioverc = PI/c_int; + DP twopioverc = 2.0*PI/c_int; + DP piovertwoc = 0.5 * pioverc; + + Vect > fmin_dlambda(Set.ntypes); + for (int n = 0; n < Set.ntypes; ++n) { + fmin_dlambda[n] = Vect (Set.epsilon[n].Npts); + for (int i = 0; i < Set.epsilon[n].Npts; ++i) + fmin_dlambda[n][i] = Vect_DP (0.0, Set.epsilon[JSC::max(n-1, 0)].Npts); + } + + for (int n = 1; n < Set.ntypes; ++n) { + for (int i = 0; i < Set.epsilon[n].Npts; ++i) { + /* + fmin_dlambda[n][i][0] = (Set.epsilon[n-1].dlambda[0]/c_int) + * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[0] > 0.0 ? + exp(-pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[0])) + /(1.0 + exp(-twopioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[0]))) + : + exp(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[0])) + /(1.0 + exp(twopioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[0])))); + //cosh(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[0]))); + + fmin_dlambda[n][i][Set.epsilon[n-1].Npts - 1] = (Set.epsilon[n-1].dlambda[Set.epsilon[n-1].Npts - 1]/c_int) + * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[Set.epsilon[n-1].Npts - 1] > 0.0 ? + exp(-pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[Set.epsilon[n-1].Npts - 1])) + /(1.0 + exp(-twopioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[Set.epsilon[n-1].Npts - 1]))) + : + exp(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[Set.epsilon[n-1].Npts - 1])) + /(1.0 + exp(twopioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[Set.epsilon[n-1].Npts - 1])))); + ///(c_int * cosh(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[Set.epsilon[n-1].Npts - 1]))); + for (int j = 1; j < Set.epsilon[n-1].Npts - 1; ++j) { + if (Set.epsilon[n].lambda[i] >= Set.epsilon[n-1].lambda[j]) + fmin_dlambda[n][i][j] = oneoverpi * atan((exp(pioverc * (-Set.epsilon[n].lambda[i] + 0.5 * (Set.epsilon[n-1].lambda[j] + Set.epsilon[n-1].lambda[j+1]))) + - exp(pioverc * (-Set.epsilon[n].lambda[i] + 0.5 * (Set.epsilon[n-1].lambda[j] + Set.epsilon[n-1].lambda[j-1])))) + /(1.0 + exp(pioverc * (-2.0 * Set.epsilon[n].lambda[i] + Set.epsilon[n-1].lambda[j] + + 0.5 * (Set.epsilon[n-1].lambda[j-1] + Set.epsilon[n-1].lambda[j+1]))))); + else + fmin_dlambda[n][i][j] = oneoverpi * atan((exp(pioverc * (Set.epsilon[n].lambda[i] - 0.5 * (Set.epsilon[n-1].lambda[j] + Set.epsilon[n-1].lambda[j-1]))) + - exp(pioverc * (Set.epsilon[n].lambda[i] - 0.5 * (Set.epsilon[n-1].lambda[j] + Set.epsilon[n-1].lambda[j+1])))) + /(1.0 + exp(pioverc * (2.0 * Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[j] + - 0.5 * (Set.epsilon[n-1].lambda[j-1] + Set.epsilon[n-1].lambda[j+1]))))); + } + + + // Boundary points: to ensure stability, make sure that \int_{-\infty}^{\infty} fmin_dlambda(., ., lambda) = 1/2. + // For j == 0, assume lambda[j-1] = -infinity: + int j = 0; + fmin_dlambda[n][i][j] = oneoverpi * atan(exp(pioverc * (-Set.epsilon[n].lambda[i] + 0.5 * (Set.epsilon[n-1].lambda[j] + Set.epsilon[n-1].lambda[j+1])))); + //fmin_dlambda[n][i][j] = oneoverpi * atan(exp(pioverc * (-Set.epsilon[n].lambda[i] + Set.epsilon[n-1].lambda[j] + 0.5 * Set.epsilon[n-1].dlambda[j]))); + // For j == Npts - 1, assume lambda[Npts] = infinity: + j = Set.epsilon[n-1].Npts - 1; + fmin_dlambda[n][i][j] = oneoverpi * atan(exp(pioverc * (Set.epsilon[n].lambda[i] - 0.5 * (Set.epsilon[n-1].lambda[j] + Set.epsilon[n-1].lambda[j-1])))); + //fmin_dlambda[n][i][j] = oneoverpi * atan(exp(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[j] - 0.5 * Set.epsilon[n-1].dlambda[j]))); + */ + + for (int j = 0; j < Set.epsilon[n-1].Npts; ++j) + fmin_dlambda[n][i][j] = oneoverpi * atan(exp(-pioverc * fabs(Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[j])) + * 2.0 * sinh(piovertwoc * Set.epsilon[n-1].dlambda[j]) + /(1.0 + exp(-twopioverc * fabs(Set.epsilon[n].lambda[i] - Set.epsilon[n-1].lambda[j])))); + + } // for i + } // for n + + return(fmin_dlambda); +} + +//void Build_fplus_dlambda (Vect >& fplus_dlambda, const Root_Density_Set& Set, DP c_int, DP mu, DP Omega, DP kBT) +Vect > Build_fplus_dlambda (const Root_Density_Set& Set, DP c_int, DP mu, DP Omega, DP kBT) +{ + DP oneoverpi = 1.0/PI; + //DP oneoverc = 1.0/c_int; + //DP twoovernc = 2.0/c_int; + DP pioverc = PI/c_int; + DP twopioverc = 2.0*PI/c_int; + DP piovertwoc = 0.5 * pioverc; + + Vect > fplus_dlambda(Set.ntypes); + for (int n = 0; n < Set.ntypes; ++n) { + fplus_dlambda[n] = Vect (Set.epsilon[n].Npts); + for (int i = 0; i < Set.epsilon[n].Npts; ++i) + fplus_dlambda[n][i] = Vect_DP (0.0, Set.epsilon[JSC::min(n+1, Set.ntypes - 1)].Npts); + } + + for (int n = 0; n < Set.ntypes - 1; ++n) { + for (int i = 0; i < Set.epsilon[n].Npts; ++i) { + /* + fplus_dlambda[n][i][0] = (Set.epsilon[n+1].dlambda[0]/c_int) + * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[0] > 0.0 ? + exp(-pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[0])) + /(1.0 + exp(-twopioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[0]))) + : + exp(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[0])) + /(1.0 + exp(twopioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[0])))); + //cosh(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[0]))); + fplus_dlambda[n][i][Set.epsilon[n+1].Npts - 1] = (Set.epsilon[n+1].dlambda[Set.epsilon[n+1].Npts - 1]/c_int) + * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[Set.epsilon[n+1].Npts - 1] > 0.0 ? + exp(-pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[Set.epsilon[n+1].Npts - 1])) + /(1.0 + exp(-twopioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[Set.epsilon[n+1].Npts - 1]))) + : + exp(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[Set.epsilon[n+1].Npts - 1])) + /(1.0 + exp(twopioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[Set.epsilon[n+1].Npts - 1])))); + ///(c_int * cosh(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[Set.epsilon[n+1].Npts - 1]))); + for (int j = 1; j < Set.epsilon[n+1].Npts - 1; ++j) { + if (Set.epsilon[n].lambda[i] >= Set.epsilon[n+1].lambda[j]) + fplus_dlambda[n][i][j] = oneoverpi * atan((exp(pioverc * (-Set.epsilon[n].lambda[i] + 0.5 * (Set.epsilon[n+1].lambda[j] + Set.epsilon[n+1].lambda[j+1]))) + - exp(pioverc * (-Set.epsilon[n].lambda[i] + 0.5 * (Set.epsilon[n+1].lambda[j] + Set.epsilon[n+1].lambda[j-1])))) + /(1.0 + exp(pioverc * (-2.0 * Set.epsilon[n].lambda[i] + Set.epsilon[n+1].lambda[j] + + 0.5 * (Set.epsilon[n+1].lambda[j-1] + Set.epsilon[n+1].lambda[j+1]))))); + else + fplus_dlambda[n][i][j] = oneoverpi * atan((exp(pioverc * (Set.epsilon[n].lambda[i] - 0.5 * (Set.epsilon[n+1].lambda[j] + Set.epsilon[n+1].lambda[j-1]))) + - exp(pioverc * (Set.epsilon[n].lambda[i] - 0.5 * (Set.epsilon[n+1].lambda[j] + Set.epsilon[n+1].lambda[j+1])))) + /(1.0 + exp(pioverc * (2.0 * Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[j] + - 0.5 * (Set.epsilon[n+1].lambda[j-1] + Set.epsilon[n+1].lambda[j+1]))))); + } + + // Boundary points: to ensure stability, make sure that \int_{-\infty}^{\infty} fmin_dlambda(., ., lambda) = 1/2. + // For j == 0, assume lambda[j-1] = -infinity: + int j = 0; + fplus_dlambda[n][i][j] = oneoverpi * atan(exp(pioverc * (-Set.epsilon[n].lambda[i] + 0.5 * (Set.epsilon[n+1].lambda[j] + Set.epsilon[n+1].lambda[j+1])))); + //fplus_dlambda[n][i][j] = oneoverpi * atan(exp(pioverc * (-Set.epsilon[n].lambda[i] + Set.epsilon[n+1].lambda[j] + 0.5 * Set.epsilon[n+1].dlambda[j]))); + // For j == Npts - 1, assume lambda[Npts] = infinity: + j = Set.epsilon[n+1].Npts - 1; + fplus_dlambda[n][i][j] = oneoverpi * atan(exp(pioverc * (Set.epsilon[n].lambda[i] - 0.5 * (Set.epsilon[n+1].lambda[j] + Set.epsilon[n+1].lambda[j-1])))); + //fplus_dlambda[n][i][j] = oneoverpi * atan(exp(pioverc * (Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[j] - 0.5 * Set.epsilon[n+1].dlambda[j]))); + */ + for (int j = 0; j < Set.epsilon[n+1].Npts; ++j) + fplus_dlambda[n][i][j] = oneoverpi * atan(exp(-pioverc * fabs(Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[j])) + * 2.0 * sinh(piovertwoc * Set.epsilon[n+1].dlambda[j]) + /(1.0 + exp(-twopioverc * fabs(Set.epsilon[n].lambda[i] - Set.epsilon[n+1].lambda[j])))); + } + } + + return(fplus_dlambda); +} + +//void Solve_2CBG_TBAE_via_refinements (Root_Density_Set& TBA_Set, DP c_int, DP mu, DP Omega, DP kBT, int Max_Secs, ofstream& LOG_outfile) +Root_Density_Set Solve_2CBG_TBAE_via_refinements (DP c_int, DP mu, DP Omega, DP kBT, int Max_Secs, ofstream& LOG_outfile, bool Save_data) +{ + // This solves the 2CBG TBAE as best as possible given the time constraint. + + clock_t StartTime = clock(); + + //int maxcycles = 5; + + int Max_CPU_ticks = 98 * (Max_Secs - 0) * CLOCKS_PER_SEC/100; // give 30 seconds to wrap up, assume we time to 2% accuracy. + + // Set basic precision needed: + DP running_prec = 1.0; + + DP refine_fraction = 0.5; // value fraction of points to be refined + + // Set basic number of types needed: + int ntypes_needed = int(kBT * log(kBT/1.0e-14)/Omega); + int ntypes = JSC::max(ntypes_needed, 1); + ntypes = JSC::min(ntypes, 10); + + //cout << "ntypes = " << ntypes << endl; + + if (Save_data) + if (ntypes >= 10) LOG_outfile << "WARNING: ntypes needs to be quite high for c_int = " << c_int << " mu = " << mu << " Omega = " << Omega + << " kBT = " << kBT << ". Set to " << ntypes << ", ideally needed: " + << ntypes_needed << ". Accuracy might be incorrectly evaluated." << endl; + + DP lambdamax = 10.0 + sqrt(JSC::max(1.0, kBT * 36.0 + mu + Omega)); // such that exp(-(lambdamax^2 - mu - Omega)/T) <~ machine_eps + int Npts = 50; + //int Npts = JSC::max(200, int(2 * lambdamax)); // For stability, we need *all* dlambda < 2. Choose max(dlambda) = 1 here for safety. + //if (Npts < int(2 * lambdamax)) + //LOG_outfile << "WARNING: dlambda = " << 2.0 * lambdamax/Npts << " might be too large to ensure numerical stability of iterations." << endl; + Vect_INT Npts_init(Npts, ntypes); + //Vect_DP lambdamax_init(c_int * 10.0, ntypes); // such that exp(-pi *lambdamax/c) <~ machine_eps + // We let the number of points fall off with increasing level n: + //for (int n = 1; n < ntypes; ++n) Npts_init[n] = JSC::max(100, 2*int(Npts/(n+1.0))); // DON'T !! Unstable. + if (Save_data) LOG_outfile << "Npts (basic) set to " << Npts_init << endl; + + Vect_DP lambdamax_init(lambdamax, ntypes); // such that exp(-pi *lambdamax/c) <~ machine_eps + Npts_init[0] = 1 * Npts; // give more precision to lowest level + lambdamax_init[0] = 10.0 + sqrt(JSC::max(1.0, kBT * 36.0 + mu + Omega)); // such that exp(-(lambdamax^2 - mu - Omega)/T) <~ machine_eps + Root_Density_Set TBA_Set (ntypes, Npts_init, lambdamax_init); + //TBA_Set = Root_Density_Set(ntypes, Npts, lambdamax); + + // Set the asymptotics of the TBA_fns: + + Set_2CBG_Asymptotics (TBA_Set, mu, Omega, kBT); + + // Initiate the functions: + + Initiate_2CBG_TBA_Functions (TBA_Set, mu, Omega); + + clock_t StopTime = clock(); + clock_t Cycle_StartTime, Cycle_StopTime; + + int CPU_ticks = StopTime - StartTime; + + int ncycles = 0; + int niter_tot = 0; + + do { + + StartTime = clock(); + + Cycle_StartTime = clock(); + + //TBA_Set.Save("Test1.dat"); + + // The running precision is an estimate of the accuracy of the free energy integral. + // Refine... returns sum_delta_tni_dlambda, so running prec is estimated as... + running_prec = Refine_2CBG_Set (TBA_Set, c_int, mu, Omega, kBT, refine_fraction); + + //TBA_Set.Save("Test2.dat"); + + //cout << "Waiting..." << endl; + //char a; cin >> a; + + Vect > a_n_dlambda = Build_a_n_dlambda (TBA_Set, c_int, mu, Omega, kBT); + Vect > fmin_dlambda = Build_fmin_dlambda (TBA_Set, c_int, mu, Omega, kBT); + Vect > fplus_dlambda = Build_fplus_dlambda (TBA_Set, c_int, mu, Omega, kBT); + + StopTime = clock(); + + CPU_ticks += StopTime - StartTime; + + int niter = 0; + int niter_max = ncycles == 0 ? 300 : 300; + + // For extrapolations: + Vect IterSet(4); + + do { + + //TBA_Set.Save("Test2.dat"); + + StartTime = clock(); + //if (ncycles <= 100) { + //if (ncycles <= 0) { + if (niter <= 10 || niter > 100) { + //Iterate_2CBG_TBAE (TBA_Set, c_int, mu, Omega, kBT); + Iterate_2CBG_TBAE (TBA_Set, a_n_dlambda, fmin_dlambda, fplus_dlambda, c_int, mu, Omega, kBT); + //cout << "iteration " << niter << "\tepsilon[0](0) = " << TBA_Set.epsilon[0].value[TBA_Set.epsilon[0].Npts/2] + // << "\tdelta epsilon[0](0) = " << TBA_Set.epsilon[0].value[TBA_Set.epsilon[0].Npts/2] - TBA_Set.epsilon[0].prev_value[TBA_Set.epsilon[0].Npts/2] + // << "\tdiff[0] = " << TBA_Set.epsilon[0].diff << "\tdiff[1] = " << TBA_Set.epsilon[1].diff + // << "\tdiff[ntypes - 1] = " << TBA_Set.epsilon[TBA_Set.ntypes - 1].diff << "\tdiff = " << TBA_Set.diff << endl; + niter++; + } + else { + Iterate_and_Extrapolate_2CBG_TBAE (TBA_Set, IterSet, a_n_dlambda, fmin_dlambda, fplus_dlambda, c_int, mu, Omega, kBT); + //cout << "extrap it " << niter << "\tepsilon[0](0) = " << TBA_Set.epsilon[0].value[TBA_Set.epsilon[0].Npts/2] + // << "\tdelta epsilon[0](0) = " + // << TBA_Set.epsilon[0].value[TBA_Set.epsilon[0].Npts/2] - TBA_Set.epsilon[0].prev_value[TBA_Set.epsilon[0].Npts/2] + // << "\tdiff = " << TBA_Set.diff << endl; + niter += 6; + } + + //TBA_Set.Save("Test3.dat"); + //cout << "Niter = " << niter << "\tdiff = " << TBA_Set.diff << "\trunning_prec " << running_prec; + //for (int n = 0; n < TBA_Set.ntypes; ++n) cout << "\t" << TBA_Set.epsilon[n].diff; + //cout << endl; + //if (ncycles >= 0) { char b; cin >> b;} + + StopTime = clock(); + CPU_ticks += StopTime - StartTime; + //} while (niter < 20 || niter < niter_max && TBA_Set.diff > running_prec && CPU_ticks < Max_CPU_ticks); + //} while (niter < 20 || niter < niter_max && TBA_Set.diff > 1.0e-10 * TBA_Set.Npts_total && CPU_ticks < Max_CPU_ticks); + } while (niter < 5 || niter < niter_max && TBA_Set.diff > 0.1 * running_prec && CPU_ticks < Max_CPU_ticks); + + if (Save_data) { + LOG_outfile << "ncycles = " << ncycles << "\trunning_prec = " << running_prec << "\t niter = " << niter + << "\tntypes = " << TBA_Set.ntypes << "\tNpts "; + for (int n = 0; n < TBA_Set.ntypes; ++n) LOG_outfile << TBA_Set.epsilon[n].Npts << " "; + //for (int n = 0; n < TBA_Set.ntypes; ++n) LOG_outfile << TBA_Set.epsilon[n].value_infty << " "; + LOG_outfile << "\tNpts_total = " << TBA_Set.Npts_total << endl + << "\tdiff = " << TBA_Set.diff << "\tGSE = " << Calculate_Gibbs_Free_Energy (TBA_Set, c_int, mu, Omega, kBT) << endl; + } + + ncycles++; + niter_tot += niter; + + if (niter == niter_max) { + if (Save_data) LOG_outfile << "Not able to improve functions enough after " << niter_max << " iterations." << endl; + //break; + } + + Cycle_StopTime = clock(); + + } while (CPU_ticks < Max_CPU_ticks - 2.0*(Cycle_StopTime - Cycle_StartTime)); + // Allow a new cycle only if there is time, assuming new cycle time < 2* last one + + if (Save_data) { + LOG_outfile << "c_int " << c_int << "\tmu " << mu << "\tOmega " << Omega << "\tkBT " << kBT + << "\tncycles = " << ncycles << "\trunning_prec = " << running_prec << "\t niter_tot = " << niter_tot + << "\tntypes = " << TBA_Set.ntypes << "\tdiff = " << TBA_Set.diff << endl << "\tNpts "; + for (int n = 0; n < TBA_Set.ntypes; ++n) LOG_outfile << TBA_Set.epsilon[n].Npts << " "; + //for (int n = 0; n < TBA_Set.ntypes; ++n) LOG_outfile << TBA_Set.epsilon[n].value_infty << " "; + LOG_outfile << "\tNpts_total = " << TBA_Set.Npts_total << endl; + } + + return(TBA_Set); + //return; +} + + +// Iterative procedures for deps/dmu or /dOmega: +void Iterate_2CBG_deps_dchempot (int option, Root_Density_Set& DSet, const Root_Density_Set& Set, + Vect >& a_n_dlambda, Vect >& fmin_dlambda, + Vect >& fplus_dlambda, DP c_int, DP mu, DP Omega, DP kBT) +{ + // Produces a new Root_Density_Set for depsilon/dmu (option == 0) or depsilon/dOmega (option == 1) from a previous iteration. + // Does NOT add types or change Npts, lambdamax values. + + //DP oneoverc = 1.0/c_int; + //DP twoovernc = 2.0/c_int; + + // First define some useful functions: + Vect depsover1plusemineps(Set.ntypes); + Vect depsover1plusepluseps(Set.ntypes); + Vect_DP depsover1pluseminepsinfty(Set.ntypes); + Vect_DP depsover1pluseplusepsinfty(Set.ntypes); + + for (int n = 0; n < Set.ntypes; ++n) { + + depsover1plusemineps[n] = Vect_DP (0.0, Set.epsilon[n].Npts); + depsover1plusepluseps[n] = Vect_DP (0.0, Set.epsilon[n].Npts); + + for (int i = 0; i < Set.epsilon[n].Npts; ++i) { + depsover1plusemineps[n][i] = Set.epsilon[n].value[i] > 0.0 ? + DSet.epsilon[n].value[i]/(1.0 + exp(-Set.epsilon[n].value[i]/kBT)) : + DSet.epsilon[n].value[i] * exp(Set.epsilon[n].value[i]/kBT)/(1.0 + exp(Set.epsilon[n].value[i]/kBT)); + depsover1plusepluseps[n][i] = Set.epsilon[n].value[i] > 0.0 ? + DSet.epsilon[n].value[i] * exp(-Set.epsilon[n].value[i]/kBT)/(1.0 + exp(-Set.epsilon[n].value[i]/kBT)) : + DSet.epsilon[n].value[i]/(1.0 + exp(Set.epsilon[n].value[i]/kBT)); + + // Keep previous rapidities: + DSet.epsilon[n].prev_value[i] = DSet.epsilon[n].value[i]; + + } + depsover1pluseminepsinfty[n] = DSet.epsilon[n].value_infty/(1.0 + exp(-Set.epsilon[n].value_infty/kBT)); + depsover1pluseplusepsinfty[n] = DSet.epsilon[n].value_infty * exp(-Set.epsilon[n].value_infty/kBT)/(1.0 + exp(-Set.epsilon[n].value_infty/kBT)); + } + + //cout << "Here 0" << endl; + + // Now do the necessary convolutions for epsilon == epsilon[0]. + // For each value of lambda, do the convolutions: + // Careful: the lambda's used for lambda (index i) are those of epsilon[0], the lambda' (index j) are for epsilon[n] !! + Vect a_n_depsover_conv(Set.ntypes); + for (int n = 0; n < Set.ntypes; ++n) { + a_n_depsover_conv[n] = Vect_DP (0.0, Set.epsilon[0].Npts); + Vect_DP f(0.0, Set.epsilon[n].Npts); + + for (int i = 0; i < Set.epsilon[0].Npts; ++i) { + a_n_depsover_conv[n][i] = 0.0; + + for (int j = 0; j < Set.epsilon[n].Npts; ++j) { + f[j] = depsover1plusepluseps[n][j] * a_n_dlambda[i][n][j]; + a_n_depsover_conv[n][i] += f[j]; + } + } + } // for (int n... We now have all the a_n * deps... at our disposal. + + // For n > nmax sum in RHS of BE for epsilon, assuming epsilon_n = epsilon_n^\infty in those cases: + // Remember: nmax = Set.ntypes - 1 + DP Smaxsum = option == 0 ? 0.0 : 2.0 * ((Set.ntypes + 1.0) * exp(-2.0 * (Set.ntypes + 1.0) * Omega/kBT)/(1.0 - exp(-2.0 * (Set.ntypes + 1.0) * Omega/kBT)) + - Set.ntypes * exp(-2.0 * Set.ntypes * Omega/kBT)/(1.0 - exp(-2.0 * Set.ntypes * Omega/kBT))); + + // Reconstruct the epsilon[0] function: + for (int i = 0; i < DSet.epsilon[0].Npts; ++i) { + DSet.epsilon[0].value[i] = -1.0; + + // Add the convolutions: + for (int n = 0; n < Set.ntypes; ++n) + DSet.epsilon[0].value[i] += a_n_depsover_conv[n][i]; + // Add the asymptotic parts of convolutions: n == 0 part is zero because of 1 + exp[epsilon[0] ] in denominator + for (int n = 1; n < Set.ntypes; ++n) + DSet.epsilon[0].value[i] += depsover1pluseplusepsinfty[n] + * (1.0 - (atan((DSet.epsilon[n].lambdamax - DSet.epsilon[0].lambda[i])/(0.5 * n * c_int)) + + atan((DSet.epsilon[n].lambdamax + DSet.epsilon[0].lambda[i])/(0.5 * n * c_int)))/PI); + + // Add the leftover summation for species n > nmax, assuming epsilon_n = epsilon_n^\infty in those cases: + DSet.epsilon[0].value[i] -= Smaxsum; + /* + if (DSet.epsilon[0].value[i] >= 0.0) { + cout << "Warning: eps[0][" << i << "] >= 0.0, " << DSet.epsilon[0].prev_value[i] << "\t" << DSet.epsilon[0].value[i] << endl; + DSet.epsilon[0].value[i] = DSet.epsilon[0].prev_value[i]; + } + else + */ + // Include some damping: + DSet.epsilon[0].value[i] = 0.1 * DSet.epsilon[0].prev_value[i] + 0.9 * DSet.epsilon[0].value[i]; + // Force boundary values to asymptotes: force boundary 10 points on each side + if (i < 10) + DSet.epsilon[0].value[i] = (1.0 - 0.1 * i) * DSet.epsilon[0].value_infty + 0.1 * i * DSet.epsilon[0].value[i]; + if (i > DSet.epsilon[0].Npts - 11) + DSet.epsilon[0].value[i] = (1.0 - 0.1 * (DSet.epsilon[0].Npts-1 - i)) * DSet.epsilon[0].value_infty + 0.1 * (DSet.epsilon[0].Npts-1 - i) * DSet.epsilon[0].value[i]; + + } + // epsilon[0] is now fully iterated. + + // Now do the remaining epsilons: + + for (int n = 1; n < Set.ntypes; ++n) { + + Vect_DP f_depsover_conv_min (0.0, Set.epsilon[n].Npts); // 'down' convolution + Vect_DP f_depsover_conv_plus (0.0, Set.epsilon[n].Npts); // 'up' convolution + + // For n = ntypes, need: + //DP em2OoT = exp(-2.0 * Omega/kBT); + //int ntypes = DSet.ntypes; + //DP depsover1pluseepsntypesinfty = option == 0 ? 0.0 : 2.0 * (1.0 - pow(em2OoT, ntypes+1.0)) + //* (ntypes * (1.0 - pow(em2OoT, ntypes+2.0)) - (ntypes + 2.0) * em2OoT * (1.0 - pow(em2OoT, DP(ntypes)))) + ///((1.0 - em2OoT) * (1.0 - pow(em2OoT, DP(ntypes))) * (1.0 - pow(em2OoT, ntypes + 2.0))) MISSING 1/1+exp...part + + Vect_DP fmin(0.0, Set.epsilon[n-1].Npts); + Vect_DP fplus(0.0, Set.epsilon[JSC::min(n+1, Set.ntypes - 1)].Npts); + + for (int i = 0; i < Set.epsilon[n].Npts; ++i) { + f_depsover_conv_min[i] = 0.0; + f_depsover_conv_plus[i] = 0.0; + + // 'down' convolutions + + if (n == 1) { + + for (int j = 0; j < Set.epsilon[0].Npts; ++j) { + fmin[j] = depsover1plusepluseps[0][j] + * fmin_dlambda[n][i][j]; + f_depsover_conv_min[i] -= fmin[j]; // Careful ! - sign here + } + //if (i == 0 || i == 1) + //cout << "i = " << i << ": " << endl << endl << fmin << endl << endl + // << depsover1plusepluseps[0] << endl << endl << fmin_dlambda[n][i] + //if (i < 10) cout << "Sum of fmin_dlambda[n][" << i << "]: " << fmin_dlambda[n][i].sum() << endl; + } // if (n == 1) + + else { // if (n != 1) + + for (int j = 0; j < Set.epsilon[n - 1].Npts; ++j) { + fmin[j] = (depsover1plusemineps[n-1][j] - depsover1pluseminepsinfty[n-1]) + * fmin_dlambda[n][i][j]; + f_depsover_conv_min[i] += fmin[j]; + } + } // if (n != 1) + + + // 'up' convolutions + if (n < Set.ntypes - 1) { + + for (int j = 0; j < Set.epsilon[n+1].Npts; ++j) { + fplus[j] = (depsover1plusemineps[n+1][j] - depsover1pluseminepsinfty[n+1]) + * fplus_dlambda[n][i][j]; + f_depsover_conv_plus[i] += fplus[j]; + } + + } // if (n < Set.ntypes - 1... + + // otherwise, we put the function to 1/2 times depsover... of epsilon[n+1] at infinity, minus the same: + else f_depsover_conv_plus[i] = 0.0; + + //Set.epsilon[n].value[i] = Set.epsilon[n].value_infty + f_Tln_conv_min[i] + f_Tln_conv_plus[i]; + // Do some damping: + DSet.epsilon[n].value[i] = 0.1 * DSet.epsilon[n].prev_value[i] + + 0.9 * (DSet.epsilon[n].value_infty + f_depsover_conv_min[i] + f_depsover_conv_plus[i]); + // Force boundary values to asymptotes: force boundary 10 points on each side + if (i < 10) + DSet.epsilon[n].value[i] = (1.0 - 0.1 * i) * DSet.epsilon[n].value_infty + 0.1 * i * DSet.epsilon[n].value[i]; + if (i > DSet.epsilon[n].Npts - 11) + DSet.epsilon[n].value[i] = (1.0 - 0.1 * (DSet.epsilon[n].Npts-1 - i)) * DSet.epsilon[n].value_infty + 0.1 * (DSet.epsilon[n].Npts-1 - i) * DSet.epsilon[n].value[i]; + + // if ((n == Set.ntypes - 1 || n == Set.ntypes - 2) && i == 0) { + /* + if (i == 0 && n < 2) { + cout << "epsilon[" << n << "][0]: " << DSet.epsilon[n].value[i] << "\t" << DSet.epsilon[n].prev_value[i] << "\t" + << DSet.epsilon[n].value_infty << "\t" << f_depsover_conv_min[i] << "\t" << f_depsover_conv_plus[i] + //<< "\tepsilon[" << n << "][1]: " << DSet.epsilon[n].value[1] << "\t" << DSet.epsilon[n].prev_value[1] << "\t" + // << DSet.epsilon[n].value_infty << "\t" << f_depsover_conv_min[1] << "\t" << f_depsover_conv_plus[1] + << "\tepsilon[" << n << "][10]: " << DSet.epsilon[n].value[10] << "\t" << DSet.epsilon[n].prev_value[10] << "\t" + << DSet.epsilon[n].value_infty << "\t" << f_depsover_conv_min[10] << "\t" << f_depsover_conv_plus[10] + << endl; + } + */ + //if (i == 0) cout << "Check: level " << n << " value_infty = " << DSet.epsilon[n].value_infty << endl; + + } // for (int i = 0... + + } // for (int n = 1... + + // All functions have now been iterated. + + // Now calculate diff: + + DSet.diff = 0.0; + + for (int n = 0; n < DSet.ntypes; ++n) { + DSet.epsilon[n].diff = 0.0; + //sum_N += Set.epsilon[n].Npts; + for (int i = 0; i < DSet.epsilon[n].Npts; ++i) { + //Set.epsilon[n].diff += pow((Set.epsilon[n].value[i] - Set.epsilon[n].prev_value[i]) + // /JSC::max(1.0, fabs(Set.epsilon[n].value[i] + Set.epsilon[n].prev_value[i])), 2.0); + //Set.epsilon[n].diff += fabs((Set.epsilon[n].value[i] - Set.epsilon[n].prev_value[i]) + // /JSC::max(1.0, fabs(Set.epsilon[n].value[i] + Set.epsilon[n].prev_value[i]))); + //DSet.epsilon[n].diff += fabs(((DSet.epsilon[n].value[i] - DSet.epsilon[n].prev_value[i])/(DSet.epsilon[n].value[i] + DSet.epsilon[n].prev_value[i])) + // * depsover1plusepluseps[n][i]); + DSet.epsilon[n].diff += DSet.epsilon[n].dlambda[i] * + fabs((DSet.epsilon[n].value[i] - DSet.epsilon[n].prev_value[i]) * depsover1plusepluseps[n][i]); + } + //DSet.epsilon[n].diff /= DSet.epsilon[n].Npts; + DSet.diff += DSet.epsilon[n].diff; + //cout << n << " " << Set.epsilon[n].diff << "\t"; + } + //DSet.diff /= Set.ntypes; + //cout << endl; + + return; +} + +/* IMPROVED VERSION BELOW +Root_Density_Set Solve_2CBG_deps_dchempot (int option, const Root_Density_Set& TBA_Set, DP c_int, DP mu, DP Omega, DP kBT, + int Max_Secs, ofstream& LOG_outfile, bool Save_data) +{ + // This solves the 2CBG deps/dmu (option == 0) or deps/dOmega (option == 1). + + clock_t StartTime = clock(); + + int Max_CPU_ticks = 98 * (Max_Secs - 10) * CLOCKS_PER_SEC/100; // give 30 seconds to wrap up, assume we time to 2% accuracy. + + // Set basic precision needed: + DP running_prec = TBA_Set.diff; + + Root_Density_Set DSet = TBA_Set; // use same number of functions and points + + // Set the asymptotics of the TBA_fns: + + Set_2CBG_deps_dchempot_Asymptotics (option, TBA_Set, DSet, mu, Omega, kBT); + + // Initiate the functions: + + Initiate_2CBG_deps_dchempot_Functions (DSet); + + Vect > a_n_dlambda = Build_a_n_dlambda (TBA_Set, c_int, mu, Omega, kBT); + Vect > fmin_dlambda = Build_fmin_dlambda (TBA_Set, c_int, mu, Omega, kBT); + Vect > fplus_dlambda = Build_fplus_dlambda (TBA_Set, c_int, mu, Omega, kBT); + + clock_t StopTime = clock(); + + int CPU_ticks = StopTime - StartTime; + + int niter = 0; + int niter_min = 20; + int niter_max = 2000; + + do { + StartTime = clock(); + Iterate_2CBG_deps_dchempot (option, DSet, TBA_Set, a_n_dlambda, fmin_dlambda, fplus_dlambda, c_int, mu, Omega, kBT); + niter++; + StopTime = clock(); + CPU_ticks += StopTime - StartTime; + //} while (niter < niter_min || niter < niter_max && DSet.diff > running_prec && CPU_ticks < Max_CPU_ticks); + } while (CPU_ticks < Max_CPU_ticks); // Just do as many iterations as allowed by time. + + if (Save_data) { + LOG_outfile << "deps_dchempot, option == " << option << endl; + LOG_outfile << "c_int " << c_int << "\tmu " << mu << "\tOmega " << Omega << "\tkBT " << kBT + << "\trunning_prec " << running_prec << " niter_tot " << niter + << "\tntypes " << DSet.ntypes << "\tdiff " << DSet.diff << endl; + } + return(DSet); +} +*/ +Root_Density_Set Solve_2CBG_deps_dchempot (int option, Root_Density_Set& TBA_Set, DP c_int, DP mu, DP Omega, DP kBT, + int Max_Secs, ofstream& LOG_outfile, bool Save_data) +{ + // This solves the 2CBG deps/dmu (option == 0) or deps/dOmega (option == 1). + + clock_t StartTime, StopTime; + + int Max_CPU_ticks = 98 * (Max_Secs - 10) * CLOCKS_PER_SEC/100; // give 30 seconds to wrap up, assume we time to 2% accuracy. + + // Set basic precision needed: + DP running_prec = TBA_Set.diff; + + // We start by converging simplified sets, with fewer points: + + //TBA_Set.Save("TBA_0.dat"); + + Root_Density_Set TBA_Set_comp = TBA_Set.Return_Compressed_and_Matched_Set(1.0); + + //TBA_Set_comp.Save("TBA_comp.dat"); + //TBA_Set.Save("TBA_1.dat"); + + Root_Density_Set DSet = TBA_Set; // this will be the final function return + Root_Density_Set DSet_comp = TBA_Set_comp; + + // Set the asymptotics of the TBA_fns: + + //DSet_comp.Save("comp_0.dat"); + + Set_2CBG_deps_dchempot_Asymptotics (option, TBA_Set_comp, DSet_comp, mu, Omega, kBT); + Set_2CBG_deps_dchempot_Asymptotics (option, TBA_Set, DSet, mu, Omega, kBT); + + // Now, start by 'converging' the comp set: + + // Initiate the functions: + + //DSet_comp.Save("comp_1.dat"); + + Initiate_2CBG_deps_dchempot_Functions (DSet_comp); + + //DSet_comp.Save("comp_2.dat"); + + Vect > a_n_dlambda_comp = Build_a_n_dlambda (TBA_Set_comp, c_int, mu, Omega, kBT); + Vect > fmin_dlambda_comp = Build_fmin_dlambda (TBA_Set_comp, c_int, mu, Omega, kBT); + Vect > fplus_dlambda_comp = Build_fplus_dlambda (TBA_Set_comp, c_int, mu, Omega, kBT); + + //cout << "epsilon[0].Npts = " << TBA_Set.epsilon[0].Npts << endl; + + int CPU_ticks = 0; + + //DSet_comp.Save("comp_init.dat"); + + int niter_comp = 0; + + do { + + //DSet_comp.Save("comp_a.dat"); + + StartTime = clock(); + Iterate_2CBG_deps_dchempot (option, DSet_comp, TBA_Set_comp, + a_n_dlambda_comp, fmin_dlambda_comp, fplus_dlambda_comp, c_int, mu, Omega, kBT); + niter_comp++; + StopTime = clock(); + CPU_ticks += StopTime - StartTime; + + //DSet_comp.Save("comp_b.dat"); + //char a; + //cout << "Waiting for next DSet iteration..." << endl; + //cin >> a; + + //} while (niter < niter_min || niter < niter_max && DSet.diff > running_prec && CPU_ticks < Max_CPU_ticks); + //cout << niter << "\t" << DSet_eighth.diff << "\t"; + //} while (CPU_ticks < Max_CPU_ticks/2 && (DSet_comp.diff > running_prec || niter < 100)); + } while (CPU_ticks < Max_CPU_ticks/2 && (niter_comp < 100 || DSet_comp.diff > running_prec)); // use at most half the time, keep rest for later. + //cout << endl; + //cout << "c_int = " << c_int << "\tmu = " << mu << "\teighth: niter = " << niter << "\tdiff = " << DSet_eighth.diff << endl << endl; + + //DSet_comp.Save("comp_final.dat"); + + DSet.Match_Densities(DSet_comp); + + //DSet.Save("full_init.dat"); + + Vect > a_n_dlambda = Build_a_n_dlambda (TBA_Set, c_int, mu, Omega, kBT); + Vect > fmin_dlambda = Build_fmin_dlambda (TBA_Set, c_int, mu, Omega, kBT); + Vect > fplus_dlambda = Build_fplus_dlambda (TBA_Set, c_int, mu, Omega, kBT); + + //CPU_ticks = 0; + + int niter = 0; + + do { + StartTime = clock(); + Iterate_2CBG_deps_dchempot (option, DSet, TBA_Set, a_n_dlambda, fmin_dlambda, fplus_dlambda, c_int, mu, Omega, kBT); + niter++; + StopTime = clock(); + CPU_ticks += StopTime - StartTime; + //} while (niter < niter_min || niter < niter_max && DSet.diff > running_prec && CPU_ticks < Max_CPU_ticks); + //cout << niter << "\t" << DSet.diff << "\t"; + //} while (CPU_ticks < Max_CPU_ticks/2 && DSet.diff > running_prec); + } while (CPU_ticks < Max_CPU_ticks && DSet_comp.diff > 1.0e-4 * running_prec); + //cout << endl; + //cout << "c_int = " << c_int << "\tmu = " << mu << "\tfull: niter = " << niter << "\tdiff = " << DSet.diff << endl << endl; + + //DSet.Save("full_final.dat"); + + // We're done !! + + if (Save_data) { + LOG_outfile << "deps_dchempot, option == " << option << endl; + LOG_outfile << "c_int " << c_int << "\tmu " << mu << "\tOmega " << Omega << "\tkBT " << kBT + << "\trunning_prec " << running_prec << " niter_comp " << niter_comp << " niter_full " << niter + << "\tntypes " << DSet.ntypes << "\tdiff " << DSet.diff << endl; + } + + //cout << "Done with first Dset... " << endl; + //char a; + //cin >> a; + + return(DSet); +} + + +TBA_Data_2CBG Solve_2CBG_TBAE_via_refinements (DP c_int, DP mu, DP Omega, DP kBT, int Max_Secs, bool Save_data) +{ + // This solves the 2CBG TBAE as best as possible given the time constraint. + + //clock_t StartTime = clock(); + + //int Max_CPU_ticks = 98 * (Max_Secs - 10) * CLOCKS_PER_SEC/100; // give 10 seconds to wrap up, assume we time to 2% accuracy. + + stringstream TBA_stringstream; + string TBA_string; + stringstream Dmu_stringstream; + string Dmu_string; + stringstream DOmega_stringstream; + string DOmega_string; + stringstream LOG_stringstream; + string LOG_string; + stringstream GFE_stringstream; + string GFE_string; + + TBA_stringstream << "EPS_2CBG_c_" << c_int << "_mu_" << mu << "_Omega_" << Omega << "_kBT_" << kBT << "_Secs_" << Max_Secs << ".dat"; + Dmu_stringstream << "EPS_2CBG_c_" << c_int << "_mu_" << mu << "_Omega_" << Omega << "_kBT_" << kBT << "_Secs_" << Max_Secs << ".dmu"; + DOmega_stringstream << "EPS_2CBG_c_" << c_int << "_mu_" << mu << "_Omega_" << Omega << "_kBT_" << kBT << "_Secs_" << Max_Secs << ".dOm"; + LOG_stringstream << "EPS_2CBG_c_" << c_int << "_mu_" << mu << "_Omega_" << Omega << "_kBT_" << kBT << "_Secs_" << Max_Secs << ".log"; + GFE_stringstream << "GFE_2CBG_c_" << c_int << "_mu_" << mu << "_Omega_" << Omega << "_kBT_" << kBT << "_Secs_" << Max_Secs << ".dat"; + + TBA_string = TBA_stringstream.str(); + const char* TBA_Cstr = TBA_string.c_str(); + + Dmu_string = Dmu_stringstream.str(); + const char* Dmu_Cstr = Dmu_string.c_str(); + + DOmega_string = DOmega_stringstream.str(); + const char* DOmega_Cstr = DOmega_string.c_str(); + + LOG_string = LOG_stringstream.str(); + const char* LOG_Cstr = LOG_string.c_str(); + + GFE_string = GFE_stringstream.str(); + const char* GFE_Cstr = GFE_string.c_str(); + + ofstream LOG_outfile; + ofstream GFE_outfile; + + if (Save_data) { + LOG_outfile.open(LOG_Cstr); + LOG_outfile.precision(6); + + GFE_outfile.open(GFE_Cstr); + GFE_outfile.precision(16); + } + + Root_Density_Set TBA_Set = Solve_2CBG_TBAE_via_refinements (c_int, mu, Omega, kBT, Max_Secs/3, LOG_outfile, Save_data); + + if (Save_data) { + // Output the functions: + TBA_Set.Save(TBA_Cstr); + } + + Root_Density_Set DSet_dmu = Solve_2CBG_deps_dchempot (0, TBA_Set, c_int, mu, Omega, kBT, Max_Secs/3, LOG_outfile, Save_data); + if (Save_data) DSet_dmu.Save(Dmu_Cstr); + + Root_Density_Set DSet_dOmega = Solve_2CBG_deps_dchempot (1, TBA_Set, c_int, mu, Omega, kBT, Max_Secs/3, LOG_outfile, Save_data); + if (Save_data) DSet_dOmega.Save(DOmega_Cstr); + + DP f = Calculate_Gibbs_Free_Energy (TBA_Set, c_int, mu, Omega, kBT); + DP dfdmu = Calculate_dGibbs_dchempot (DSet_dmu, TBA_Set, c_int, mu, Omega, kBT); + DP dfdOm = Calculate_dGibbs_dchempot (DSet_dOmega, TBA_Set, c_int, mu, Omega, kBT); + + if (Save_data) + GFE_outfile << f << "\t" << TBA_Set.diff + << "\t" << dfdmu << "\t" << dfdOm << "\t" << 0.5 * (-dfdmu - dfdOm) << "\t" << 0.5 * (-dfdmu + dfdOm) + << endl; + + else cout << setprecision(16) << f << "\t" << 0.5 * (-dfdmu - dfdOm) << "\t" << 0.5 * (-dfdmu + dfdOm); + + if (Save_data) { + LOG_outfile.close(); + GFE_outfile.close(); + } + + TBA_Data_2CBG answer; + answer.c_int = c_int; + answer.mu = mu; + answer.Omega = Omega; + answer.kBT = kBT; + answer.f = f; + answer.n1 = 0.5 * (-dfdmu - dfdOm); + answer.n2 = 0.5 * (-dfdmu + dfdOm); + + return(answer); +} + +void GFE_muscan_2CBG (DP c_int, DP mu_min, DP mu_max, DP Omega, DP kBT, int Npts_mu, int Max_Secs, bool Save_data) +{ + DP dmu = (mu_max - mu_min)/(Npts_mu - 1); + + stringstream LOG_stringstream; + string LOG_string; + stringstream GFE_stringstream; + string GFE_string; + + LOG_stringstream << "GFE_2CBG_c_" << c_int << "_mu_min_" << mu_min << "_mu_max_" << mu_max + << "_Npts_mu_" << Npts_mu << "_Omega_" << Omega << "_kBT_" << kBT << "_Secs_" << Max_Secs << ".log"; + GFE_stringstream << "GFE_2CBG_c_" << c_int << "_mu_min_" << mu_min << "_mu_max_" << mu_max + << "_Npts_mu_" << Npts_mu << "_Omega_" << Omega << "_kBT_" << kBT << "_Secs_" << Max_Secs << ".dat"; + + LOG_string = LOG_stringstream.str(); + const char* LOG_Cstr = LOG_string.c_str(); + + GFE_string = GFE_stringstream.str(); + const char* GFE_Cstr = GFE_string.c_str(); + + ofstream LOG_outfile; + ofstream GFE_outfile; + + LOG_outfile.open(LOG_Cstr); + LOG_outfile.precision(6); + + GFE_outfile.open(GFE_Cstr); + GFE_outfile.precision(16); + + + Root_Density_Set Scan_Set (10, 10, 10.0);; + Root_Density_Set Scan_dSet_dmu (10, 10, 10.0);; + Root_Density_Set Scan_dSet_dOmega (10, 10, 10.0);; + DP mu; + int Max_Secs_per_mu_pt = Max_Secs/Npts_mu; + + for (int imu = 0; imu < Npts_mu; ++imu) { + mu = mu_min + imu * dmu; + Scan_Set = Solve_2CBG_TBAE_via_refinements (c_int, mu, Omega, kBT, Max_Secs_per_mu_pt/3, LOG_outfile, Save_data); + //Solve_2CBG_TBAE_via_refinements (Scan_Set, c_int, mu, Omega, kBT, Max_Secs_per_mu_pt, LOG_outfile); + + Scan_dSet_dmu = Solve_2CBG_deps_dchempot (0, Scan_Set, c_int, mu, Omega, kBT, Max_Secs_per_mu_pt/3, LOG_outfile, Save_data); + + Scan_dSet_dOmega = Solve_2CBG_deps_dchempot (1, Scan_Set, c_int, mu, Omega, kBT, Max_Secs_per_mu_pt/3, LOG_outfile, Save_data); + + DP dfdmu = Calculate_dGibbs_dchempot (Scan_dSet_dmu, Scan_Set, c_int, mu, Omega, kBT); + DP dfdOm = Calculate_dGibbs_dchempot (Scan_dSet_dOmega, Scan_Set, c_int, mu, Omega, kBT); + + GFE_outfile << mu << "\t" << Calculate_Gibbs_Free_Energy (Scan_Set, c_int, mu, Omega, kBT) << "\t" << Scan_Set.diff + << "\t" << dfdmu << "\t" << dfdOm << "\t" << 0.5 * (-dfdmu - dfdOm) << "\t" << 0.5 * (-dfdmu + dfdOm) + << endl; + } // for imu + + LOG_outfile.close(); + GFE_outfile.close(); + + return; +} + +} // namespace JSC diff --git a/src/TBA/TBA_LiebLin.cc b/src/TBA/TBA_LiebLin.cc new file mode 100644 index 0000000..f274539 --- /dev/null +++ b/src/TBA/TBA_LiebLin.cc @@ -0,0 +1,799 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: TBA_LiebLin.cc + +Purpose: TBA for Lieb-Liniger + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + // T = 0 functions + + DP LiebLin_a2_kernel (DP c_int, DP lambdadiff) + { + return(c_int/(PI * (lambdadiff * lambdadiff + c_int * c_int))); + } + + void Iterate_LiebLin_rho_GS (DP c_int, DP k_F, Root_Density& rho_GS) + { + // Performs an iteration of the TBA equation for the dressed charge: + + // First, copy the actual values into the previous ones: + for (int i = 0; i < rho_GS.Npts; ++i) rho_GS.prev_value[i] = rho_GS.value[i]; + + // Calculate the new values: + DP conv = 0.0; + for (int i = 0; i < rho_GS.Npts; ++i) { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < rho_GS.Npts; ++j) + conv += fabs(rho_GS.lambda[j]) > k_F ? 0.0 : + LiebLin_a2_kernel (c_int, rho_GS.lambda[i] - rho_GS.lambda[j]) + * rho_GS.prev_value[j] * rho_GS.dlambda[j]; + rho_GS.value[i] = fabs(rho_GS.lambda[i]) < k_F ? 1.0/twoPI + conv : 0.0; + } + + // Calculate the sum of differences: + rho_GS.diff = 0.0; + for (int i = 0; i < rho_GS.Npts; ++i) + rho_GS.diff += fabs(rho_GS.value[i] - rho_GS.prev_value[i]); + + return; + } + + Root_Density LiebLin_rho_GS (DP c_int, DP k_F, DP lambdamax, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the ground state + // density function of the Lieb-Liniger model. + + Root_Density rho_GS(Npts, lambdamax); + + // We now attempt to find a solution... + + do { + Iterate_LiebLin_rho_GS (c_int, k_F, rho_GS); + } while (rho_GS.diff > req_prec); + + return(rho_GS); + } + + DP Density_GS (Root_Density& rho_GS) + { + DP n = 0.0; + for (int i = 0; i < rho_GS.Npts; ++i) n += rho_GS.value[i] * rho_GS.dlambda[i]; + + return(n); + } + + DP k_F_given_n (DP c_int, DP n, DP lambdamax, int Npts, DP req_prec) + { + DP k_F = 1.0; + DP dk_F = k_F; + Root_Density rho_GS = LiebLin_rho_GS (c_int, k_F, lambdamax, Npts, req_prec); + DP n_found = Density_GS (rho_GS); + do { + dk_F *= 0.7; + k_F += (n_found < n) ? dk_F : -dk_F; + rho_GS = LiebLin_rho_GS (c_int, k_F, lambdamax, Npts, req_prec); + n_found = Density_GS (rho_GS); + + //cout << "k_F " << k_F << "\tn_found " << n_found << endl; + + } while (fabs(dk_F) > req_prec && dk_F > 1.0/Npts + && fabs(n - n_found) > req_prec && fabs(n - n_found) > 1.0/Npts); + + return(k_F); + } + + void Iterate_LiebLin_Z_GS (DP c_int, DP k_F, Root_Density& Z_GS) + { + // Performs an iteration of the TBA equation for the dressed charge: + + // First, copy the actual values into the previous ones: + for (int i = 0; i < Z_GS.Npts; ++i) Z_GS.prev_value[i] = Z_GS.value[i]; + + // Calculate the new values: + DP conv = 0.0; + for (int i = 0; i < Z_GS.Npts; ++i) { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < Z_GS.Npts; ++j) + conv += fabs(Z_GS.lambda[j]) > k_F ? 0.0 : + LiebLin_a2_kernel (c_int, Z_GS.lambda[i] - Z_GS.lambda[j]) + * Z_GS.prev_value[j] * Z_GS.dlambda[j]; + Z_GS.value[i] = 1.0 + conv; + } + + // Calculate the sum of differences: + Z_GS.diff = 0.0; + for (int i = 0; i < Z_GS.Npts; ++i) + Z_GS.diff += fabs(Z_GS.value[i] - Z_GS.prev_value[i]); + + return; + } + + Root_Density LiebLin_Z_GS (DP c_int, DP k_F, DP lambdamax, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the ground state + // dressed charge function of the Lieb-Liniger model. + + Root_Density Z_GS(Npts, lambdamax); + + // We now attempt to find a solution... + + do { + Iterate_LiebLin_Z_GS (c_int, k_F, Z_GS); + } while (Z_GS.diff > req_prec); + + return(Z_GS); + } + + void Iterate_LiebLin_Fbackflow_GS (DP c_int, DP k_F, DP lambda, Root_Density& F_GS) + { + // Performs an iteration of the TBA equation for the backflow: + + // First, copy the actual values into the previous ones: + for (int i = 0; i < F_GS.Npts; ++i) F_GS.prev_value[i] = F_GS.value[i]; + + // Calculate the new values: + DP conv = 0.0; + for (int i = 0; i < F_GS.Npts; ++i) { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < F_GS.Npts; ++j) + conv += fabs(F_GS.lambda[j]) > k_F ? 0.0 : + LiebLin_a2_kernel (c_int, F_GS.lambda[i] - F_GS.lambda[j]) + * F_GS.prev_value[j] * F_GS.dlambda[j]; + F_GS.value[i] = 0.5 + atan((F_GS.lambda[i] - lambda)/c_int)/PI + conv; + } + + // Calculate the sum of differences: + F_GS.diff = 0.0; + for (int i = 0; i < F_GS.Npts; ++i) + F_GS.diff += fabs(F_GS.value[i] - F_GS.prev_value[i]); + + return; + } + + Root_Density LiebLin_Fbackflow_GS (DP c_int, DP k_F, DP lambdamax, DP lambda, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the ground state + // F backflow for parameter \lambda. + + Root_Density F_GS(Npts, lambdamax); + + // We now attempt to find a solution... + + do { + Iterate_LiebLin_Fbackflow_GS (c_int, k_F, lambda, F_GS); + } while (F_GS.diff > req_prec); + + return(F_GS); + } + + + + // Finite T functions: + /* + // from JSC_TBA.h + + struct LiebLin_TBA_Solution { + + DP c_int; + DP mu; + DP nbar; + DP kBT; + Root_Density epsilon; + Root_Density depsilon_dmu; + Root_Density rho; + Root_Density rhoh; + + LiebLin_TBA_Solution (DP c_int_ref, DP mu_ref, DP kBT_ref, int Npts_ref, DP req_diff, int Max_Secs); + }; + */ + + LiebLin_TBA_Solution::LiebLin_TBA_Solution (DP c_int_ref, DP mu_ref, DP kBT_ref, DP req_diff, int Max_Secs) + : c_int(c_int_ref), mu(mu_ref), kBT(kBT_ref) + { + epsilon = LiebLin_epsilon_TBA (c_int, mu, kBT, req_diff, Max_Secs); + depsilon_dmu = LiebLin_depsilon_dmu_TBA (c_int, mu, kBT, req_diff, Max_Secs, epsilon); + rho = LiebLin_rho_TBA (kBT, epsilon, depsilon_dmu); + rhoh = LiebLin_rhoh_TBA (kBT, epsilon, depsilon_dmu); + nbar = LiebLin_nbar_TBA (rho); + ebar = LiebLin_ebar_TBA (rho); + sbar = LiebLin_sbar_TBA (rho, rhoh); + } + /* + LiebLin_TBA_Solution::LiebLin_TBA_Solution (DP c_int_ref, DP mu_ref, DP kBT_ref, int Npts_ref, DP req_diff, int Max_Secs, const LiebLin_TBA_Solution& prev_sol) + : c_int(c_int_ref), mu(mu_ref), kBT(kBT_ref) + { + epsilon = LiebLin_epsilon_TBA (c_int, mu, kBT, Npts_ref, req_diff, Max_Secs, prev_sol.epsilon); + depsilon_dmu = LiebLin_depsilon_dmu_TBA (c_int, mu, kBT, Npts_ref, req_diff, Max_Secs, epsilon, prev_sol.depsilon_dmu); + rho = LiebLin_rho_TBA (kBT, epsilon, depsilon_dmu); + rhoh = LiebLin_rhoh_TBA (kBT, epsilon, depsilon_dmu); + nbar = LiebLin_nbar_TBA (rho); + } + */ + + + LiebLin_TBA_Solution LiebLin_TBA_Solution_fixed_nbar_ebar (DP c_int, DP nbar_required, DP ebar_required, DP req_diff, int Max_Secs) + { + // This function finds the TBA solution for a required nbar and ebar (mean energy). + // We here try to triangulate the temperature; the chemical potential is triangulated + // using the function LiebLin_TBA_Solution_fixed_nbar. + + // FIRST VERSION + DP lnkBT = 1.0; + DP dlnkBT = 1.0; + + LiebLin_TBA_Solution tbasol = LiebLin_TBA_Solution_fixed_nbar (c_int, nbar_required, exp(lnkBT), req_diff, Max_Secs); + DP lnkBT_prev = lnkBT; + DP ebar_prev = LiebLin_ebar_TBA (tbasol.rho); + + DP dlnkBT_prev, debardlnkBT; + + lnkBT += dlnkBT; + tbasol = LiebLin_TBA_Solution_fixed_nbar (c_int, nbar_required, exp(lnkBT), req_diff, Max_Secs); + + int niter = 1; + while (fabs(tbasol.ebar - ebar_required) > req_diff) { + + dlnkBT_prev = dlnkBT; + lnkBT_prev = lnkBT; + // Calculate a new kBT based on earlier data: + debardlnkBT = (tbasol.ebar - ebar_prev)/dlnkBT; + ebar_prev = tbasol.ebar; + dlnkBT = (ebar_required - tbasol.ebar)/debardlnkBT; + if (fabs(dlnkBT) > fabs(dlnkBT_prev)) dlnkBT *= fabs(dlnkBT_prev/dlnkBT); // make sure we don't blow up. + lnkBT += dlnkBT; + tbasol = LiebLin_TBA_Solution_fixed_nbar(c_int, nbar_required, exp(lnkBT), req_diff, Max_Secs); + niter++; + //cout << setprecision(16) << "kBT: niter " << niter << "\tebar = " << tbasol.ebar << "\trequired = " << ebar_required << "\tlnkBT = " << lnkBT << "\tdlnkBT = " << dlnkBT << endl; + } + // FIRST VERSION + + /* SECOND VERSION + DP lnkBT = 1.0; + DP dlnkBT = 1.0; + + DP mu = 2.0; + DP dmu = 1.0; + + LiebLin_TBA_Solution tbasol = LiebLin_TBA_Solution(c_int, mu, kBT, req_diff, Max_Secs); + DP mu_prev = mu; + DP nbar_prev = tbasol.nbar; + DP lnkBT_prev = lnkBT; + DP ebar_prev = LiebLin_ebar_TBA (tbasol.rho); + + DP dmu_prev, dnbardmu; + DP dlnkBT_prev, debardlnkBT; + */ // SECOND VERSION + + return(tbasol); + } + + + + LiebLin_TBA_Solution LiebLin_TBA_Solution_fixed_nbar (DP c_int, DP nbar_required, DP kBT, DP req_diff, int Max_Secs) + { + // Find the required mu. Provide some initial guess. + /* + DP gamma_required = c_int/nbar_required; + + + // We define a matrix of mu's (calculated once and for all and filled in here) + // from which we can define an accurate first guess for mu, given gamma_required and kBT. + + // The gamma's are by definition (1/16) * 2^{igamma}, + // and the kBT are (1/16) * 2^{iT}. + int ndata = 16; + Vect gamma(ndata); + Vect Tred(ndata); // reduced temperature, kBT/ + for (int i = 0; i < ndata; ++i) { + gamma[i] = 0.0625 * pow(2.0, i); + T[i] = 0.0625 * pow(2.0, i); + } + SQMat mudata(ndata); + // This data was computed separately, for unit filling, using 512 points, to accuracy 10^-4: + + // Figure out which index igamma we should use: + // (1/16) * 2^{igamma_low} <~ gamma_required, so + int igamma_below = int(log(gamma_required * 16.0)/log(2.0)); + // Similarly, + int iT_below = int(log(kBT * 16.0)/log(2.0)); + + igamma_below = JSC::max(igamma_below, 0); + igamma_below = JSC::min(igamma_below, ndata - 1); + iT_below = JSC::max(iT_below, 0); + iT_below = JSC::min(iT_below, ndata - 1); + // We use indices igamma_below, igamma_below + 1, iT_below, iT_below + 1 to guess mu: + // do a four-point extrapolation, + DP mu = ((gamma[igamma_below + 1] - gamma_required) * (T[iT_below + 1] - kBT) * mudata[igamma_below][iT_below] + + (gamma_required - gamma[igamma_below]) * (T[iT_below + 1] - kBT) * mudata[igamma_below + 1][iT_below] + + (gamma[igamma_below + 1] - gamma_required) * (kBT - T[iT_below]) * mudata[igamma_below][iT_below + 1] + + (gamma_required - gamma[igamma_below]) * (kBT - T[iT_below]) * mudata[igamma_below + 1][iT_below + 1]) + /((gamma[igamma_below + 1] - gamma[igamma_below]) * (T[iT_below + 1] - T[iT_below])); + + // Translate to the required filling: + + + DP dnbardmu = ; + */ + + DP mu = 2.0; + DP dmu = 1.0; + + LiebLin_TBA_Solution tbasol = LiebLin_TBA_Solution(c_int, mu, kBT, req_diff, Max_Secs); + DP mu_prev = mu; + DP nbar_prev = tbasol.nbar; + + DP dmu_prev, dnbardmu; + + mu += dmu; + tbasol = LiebLin_TBA_Solution(c_int, mu, kBT, req_diff, Max_Secs); + + int niter = 1; + while (fabs(tbasol.nbar - nbar_required) > req_diff) { + + dmu_prev = dmu; + mu_prev = mu; + // Calculate a new mu based on earlier data: + dnbardmu = (tbasol.nbar - nbar_prev)/dmu; + nbar_prev = tbasol.nbar; + dmu = (nbar_required - tbasol.nbar)/dnbardmu; + if (fabs(dmu) > 2.0 * fabs(dmu_prev)) dmu = 2.0*dmu * fabs(dmu_prev/dmu); // make sure we don't blow up. + mu += dmu; + tbasol = LiebLin_TBA_Solution(c_int, mu, kBT, req_diff, Max_Secs); + niter++; + //cout << setprecision(16) << "\tmu: niter " << niter << "\tnbar = " << tbasol.nbar << "\trequired = " << nbar_required<< "\tmu = " << mu << "\tdmu = " << dmu << endl; + } + + return(tbasol); + } + + + DP Refine_LiebLin_epsilon_TBA (Root_Density& epsilon, DP c_int, DP mu, DP kBT, DP refine_fraction) + { + // This function replaces Set by a new set with more points, where + // Tln(...) needs to be evaluated more precisely. + + // The return value is the max of delta_tni found. + + // First, calculate the needed Tln... + Vect Tln1plusemineps(epsilon.Npts); + + for (int i = 0; i < epsilon.Npts; ++i) { + Tln1plusemineps[i] = epsilon.value[i] > 0.0 ? + kBT * (epsilon.value[i] < 24.0 * kBT ? log(1.0 + exp(-epsilon.value[i]/kBT)) : exp(-epsilon.value[i]/kBT)) + : -epsilon.value[i] + kBT * log (1.0 + exp(epsilon.value[i]/kBT)); + } + + // Now find the achieved delta_tni + DP max_delta_tni_dlambda = 0.0; + DP sum_delta_tni_dlambda = 0.0; + + Vect tni(0.0, epsilon.Npts); + Vect tni_ex(0.0, epsilon.Npts); + + DP measure_factor = 0.0; + + for (int i = 1; i < epsilon.Npts - 1; ++i) { + measure_factor = (epsilon.value[i] > 0.0 ? exp(-epsilon.value[i]/kBT)/(1.0 + exp(-epsilon.value[i]/kBT)) + : 1.0/(1.0 + exp(epsilon.value[i]/kBT))); + + tni[i] = measure_factor * epsilon.value[i]; + tni_ex[i] = measure_factor * (epsilon.value[i-1] * (epsilon.lambda[i+1] - epsilon.lambda[i]) + + epsilon.value[i+1] * (epsilon.lambda[i] - epsilon.lambda[i-1])) + /(epsilon.lambda[i+1] - epsilon.lambda[i-1]); + + max_delta_tni_dlambda = JSC::max(max_delta_tni_dlambda, fabs(tni[i] - tni_ex[i]) * epsilon.dlambda[i]); + + sum_delta_tni_dlambda += fabs(tni[i] - tni_ex[i]) * epsilon.dlambda[i]; + } + + // We now determine the locations where we need to add points + Vect need_new_point_around(false, epsilon.Npts); + int nr_new_points_needed = 0; + + for (int i = 1; i < epsilon.Npts - 1; ++i) { + if (fabs(tni[i] - tni_ex[i]) * epsilon.dlambda[i] > (1.0 - refine_fraction) * max_delta_tni_dlambda) { + need_new_point_around[i] = true; + // Do also the symmetric ones... Require need...[n][i] = need...[n][Npts - 1 - i] + need_new_point_around[epsilon.Npts - 1 - i] = true; + } + } + for (int i = 0; i < epsilon.Npts; ++i) + if (need_new_point_around[i]) nr_new_points_needed++; + + // Now insert the new points between existing points: + // Update all data via interpolation: + Root_Density epsilon_before_update = epsilon; + epsilon = Root_Density(epsilon_before_update.Npts + nr_new_points_needed, epsilon_before_update.lambdamax); + + int nr_pts_added = 0; + for (int i = 0; i < epsilon_before_update.Npts; ++i) { + if (!need_new_point_around[i]) { + epsilon.lambda[i + nr_pts_added] = epsilon_before_update.lambda[i]; + epsilon.dlambda[i + nr_pts_added] = epsilon_before_update.dlambda[i]; + epsilon.value[i + nr_pts_added] = epsilon_before_update.value[i]; + } + else if (need_new_point_around[i]) { + epsilon.lambda[i + nr_pts_added] = epsilon_before_update.lambda[i] - 0.25 * epsilon_before_update.dlambda[i]; + epsilon.dlambda[i + nr_pts_added] = 0.5 * epsilon_before_update.dlambda[i]; + epsilon.value[i + nr_pts_added] = epsilon_before_update.Return_Value(epsilon.lambda[i + nr_pts_added]); + nr_pts_added++; + epsilon.lambda[i + nr_pts_added] = epsilon_before_update.lambda[i] + 0.25 * epsilon_before_update.dlambda[i]; + epsilon.dlambda[i + nr_pts_added] = 0.5 * epsilon_before_update.dlambda[i]; + epsilon.value[i + nr_pts_added] = epsilon_before_update.Return_Value(epsilon.lambda[i + nr_pts_added]); + } + } + if (nr_pts_added != nr_new_points_needed) { + cout << nr_pts_added << "\t" << nr_new_points_needed << endl; + JSCerror("Wrong counting of new points in Insert_new_points."); + } + + + // Check boundary values; if too different from value_infty, extend limits + if (exp(-epsilon.value[0]/kBT) > 0.001 * max_delta_tni_dlambda) { + + int nr_pts_to_add_left = epsilon.Npts/10; + int nr_pts_to_add_right = epsilon.Npts/10; + Root_Density epsilon_before_update = epsilon; + epsilon = Root_Density(epsilon_before_update.Npts + nr_pts_to_add_left + nr_pts_to_add_right, epsilon_before_update.lambdamax + nr_pts_to_add_left * epsilon_before_update.dlambda[0] + nr_pts_to_add_right * epsilon_before_update.dlambda[epsilon_before_update.Npts - 1]); + + // Initialize points added on left + for (int i = 0; i < nr_pts_to_add_left; ++i) { + epsilon.lambda[i] = epsilon_before_update.lambda[i] + (i - nr_pts_to_add_left) * epsilon_before_update.dlambda[0]; + epsilon.dlambda[i] = epsilon_before_update.dlambda[0]; + epsilon.value[i] = epsilon_before_update.value[0]; + } + // Initialize middle (unchanged) points + for (int i = 0; i < epsilon_before_update.Npts; ++i) { + epsilon.lambda[nr_pts_to_add_left + i] = epsilon_before_update.lambda[i]; + epsilon.dlambda[nr_pts_to_add_left + i] = epsilon_before_update.dlambda[i]; + epsilon.value[nr_pts_to_add_left + i] = epsilon_before_update.value[i]; + } + // Initialize points added on right + for (int i = 0; i < nr_pts_to_add_right; ++i) { + epsilon.lambda[nr_pts_to_add_left + epsilon_before_update.Npts + i] = epsilon_before_update.lambda[epsilon_before_update.Npts - 1] + (i+1) * epsilon_before_update.dlambda[epsilon_before_update.Npts - 1]; + epsilon.dlambda[nr_pts_to_add_left + epsilon_before_update.Npts + i] = epsilon_before_update.dlambda[epsilon_before_update.Npts - 1]; + epsilon.value[nr_pts_to_add_left + epsilon_before_update.Npts + i] = epsilon_before_update.value[epsilon_before_update.Npts - 1]; + } + } + + return(sum_delta_tni_dlambda); + } + + // epsilon for a given mu: + Root_Density LiebLin_epsilon_TBA (DP c_int, DP mu, DP kBT, DP req_diff, int Max_Secs) + { + + clock_t StartTime = clock(); + int Max_CPU_ticks = 98 * (Max_Secs - 0) * CLOCKS_PER_SEC/100; // give 30 seconds to wrap up, assume we time to 2% accuracy. + + // Set basic precision needed: + DP running_prec = 1.0; + + DP refine_fraction = 0.5; // value fraction of points to be refined + + DP lambdamax_init = 10.0 * sqrt(JSC::max(1.0, kBT + mu)); // such that exp(-(lambdamax^2 - mu - Omega)/T) <~ machine_eps + + int Npts = 50; + + Root_Density epsilon(Npts, lambdamax_init); + + // Initiate the function: + for (int i = 0; i < epsilon.Npts; ++i) { + epsilon.value[i] = epsilon.lambda[i] * epsilon.lambda[i] - mu; + epsilon.prev_value[i] = epsilon.value[i]; + } + + clock_t StopTime = clock(); + + int CPU_ticks = StopTime - StartTime; + + int ncycles = 0; + int niter_tot = 0; + DP previous_running_prec; + + DP oneoverpi = 1.0/PI; + DP oneoverc = 1.0/c_int; + + do { + + StartTime = clock(); + + // The running precision is an estimate of the accuracy of the free energy integral. + // Refine... returns sum_delta_tni_dlambda, so running prec is estimated as... + previous_running_prec = running_prec; + running_prec = Refine_LiebLin_epsilon_TBA (epsilon, c_int, mu, kBT, refine_fraction); + //cout << "ncycles = " << ncycles << "\trunning_prec = " << running_prec << endl; + running_prec = JSC::min(running_prec, previous_running_prec); + + // Now iterate to convergence for given discretization + int niter = 0; + int niter_max = ncycles == 0 ? 300 : 300; + + do { + + StartTime = clock(); + + // Iterate the TBA equation for epsilon: + Vect Tln1plusemineps(epsilon.Npts); + for (int i = 0; i < epsilon.Npts; ++i) { + Tln1plusemineps[i] = epsilon.value[i] > 0.0 ? + kBT * (epsilon.value[i] < 24.0 * kBT ? log(1.0 + exp(-epsilon.value[i]/kBT)) : exp(-epsilon.value[i]/kBT) * (1.0 - 0.5 * exp(-epsilon.value[i]/kBT))) + //kBT * log(1.0 + exp(-epsilon.value[i]/kBT)) + : + -epsilon.value[i] + kBT * (-epsilon.value[i] < 24.0 * kBT ? log (1.0 + exp(epsilon.value[i]/kBT)) : exp(epsilon.value[i]/kBT) * (1.0 - 0.5 * exp(epsilon.value[i]/kBT))); + //-epsilon.value[i] + kBT * log (1.0 + exp(epsilon.value[i]/kBT)); + // Keep previous rapidities: + epsilon.prev_value[i] = epsilon.value[i]; + } + + Vect a_2_Tln_conv(epsilon.Npts); + + for (int i = 0; i < epsilon.Npts; ++i) { + a_2_Tln_conv[i] = 0.0; + + for (int j = 0; j < epsilon.Npts; ++j) a_2_Tln_conv[i] += oneoverpi * (atan(oneoverc * (epsilon.lambda[i] - epsilon.lambda[j] + 0.5 * epsilon.dlambda[j])) - atan(oneoverc * (epsilon.lambda[i] - epsilon.lambda[j] - 0.5 * epsilon.dlambda[j]))) * Tln1plusemineps[j]; + + } // a_2_Tln_conv is now calculated + + // Reconstruct the epsilon function: + for (int i = 0; i < epsilon.Npts; ++i) { + epsilon.value[i] = pow(epsilon.lambda[i], 2.0) - mu; + + // Add the convolution: + epsilon.value[i] -= a_2_Tln_conv[i]; + + // Include some damping: + //epsilon.value[i] = 0.1 * epsilon.prev_value[i] + 0.9 * epsilon.value[i]; + epsilon.value[i] = 0.1 * epsilon.prev_value[i] + 0.9 * epsilon.value[i]; + //* (1.0 + (epsilon.value[i] - epsilon.prev_value[i])/fabs(epsilon.value[i] + epsilon.prev_value[i])); + } + + niter++; + // epsilon is now fully iterated. + + // Calculate the diff: + epsilon.diff = 0.0; + for (int i = 0; i < epsilon.Npts; ++i) + epsilon.diff += epsilon.dlambda[i] * + (epsilon.value[i] > 0.0 ? + exp(-epsilon.value[i]/kBT)/(1.0 + exp(-epsilon.value[i]/kBT)) : 1.0/(1.0 + exp(epsilon.value[i]/kBT))) + * fabs(epsilon.value[i] - epsilon.prev_value[i]); + + StopTime = clock(); + CPU_ticks += StopTime - StartTime; + + //cout << "epsilon: niter = " << niter << "\tdiff = " << epsilon.diff << endl; + //cout << epsilon.lambda[0] << "\t" << epsilon.dlambda[0] << endl; + //cout << "a_2_Tln_conv[0] = " << a_2_Tln_conv[0] << "\tTln1plusemineps[0] = " << Tln1plusemineps[0] << endl; + } while (niter < 5 || niter < niter_max && CPU_ticks < Max_CPU_ticks && epsilon.diff > 0.1*running_prec); + + ncycles++; + niter_tot += niter; + + //cout << "End of a cycle: niter = " << niter << "\tniter_tot = " << niter_tot << "\tepsilon.diff = " << epsilon.diff << "\tNpts = " << epsilon.Npts << "\tlambdamax = " << epsilon.lambda[0] << "\trunning_prec = " << running_prec << "\treq_diff = " << req_diff << endl; + } // do cycles + while (ncycles < 5 || running_prec > req_diff && CPU_ticks < Max_CPU_ticks); + + return(epsilon); + } + + + // depsilon/dmu for a given mu + Root_Density LiebLin_depsilon_dmu_TBA (DP c_int, DP mu, DP kBT, DP req_diff, int Max_Secs, const Root_Density& epsilon) + { + + clock_t StartTime = clock(); + int Max_CPU_ticks = 98 * (Max_Secs - 0) * CLOCKS_PER_SEC/100; // give 30 seconds to wrap up, assume we time to 2% accuracy. + + Root_Density depsilon_dmu = epsilon; + + // Initiate the functions: + for (int i = 0; i < depsilon_dmu.Npts; ++i) { + depsilon_dmu.value[i] = -1.0; + depsilon_dmu.prev_value[i] = depsilon_dmu.value[i]; + } + + clock_t StopTime = clock(); + + int CPU_ticks = StopTime - StartTime; + + int niter = 0; + int niter_max = 1000; + + DP oneoverpi = 1.0/PI; + DP oneoverc = 1.0/c_int; + + do { + + StartTime = clock(); + + // Iterate the TBA equation for depsilon_dmu: + Vect depsover1plusepluseps(depsilon_dmu.Npts); + for (int i = 0; i < depsilon_dmu.Npts; ++i) { + depsover1plusepluseps[i] = epsilon.value[i] > 0.0 ? + depsilon_dmu.value[i] * exp(-epsilon.value[i]/kBT)/(1.0 + exp(-epsilon.value[i]/kBT)) : + depsilon_dmu.value[i]/(1.0 + exp(epsilon.value[i]/kBT)); + + // Keep previous rapidities: + depsilon_dmu.prev_value[i] = depsilon_dmu.value[i]; + } + + Vect a_2_depsover1plusepluseps_conv(depsilon_dmu.Npts); + + for (int i = 0; i < depsilon_dmu.Npts; ++i) { + a_2_depsover1plusepluseps_conv[i] = 0.0; + + for (int j = 0; j < depsilon_dmu.Npts; ++j) + a_2_depsover1plusepluseps_conv[i] += oneoverpi * (atan(oneoverc * (epsilon.lambda[i] - epsilon.lambda[j] + 0.5 * epsilon.dlambda[j])) - atan(oneoverc * (epsilon.lambda[i] - epsilon.lambda[j] - 0.5 * epsilon.dlambda[j]))) * depsover1plusepluseps[j]; + } + + // Reconstruct the depsilon_dmu function: + for (int i = 0; i < epsilon.Npts; ++i) { + depsilon_dmu.value[i] = -1.0; + + // Add the convolution: + depsilon_dmu.value[i] += a_2_depsover1plusepluseps_conv[i]; + + // Include some damping: + depsilon_dmu.value[i] = 0.1 * depsilon_dmu.prev_value[i] + 0.9 * depsilon_dmu.value[i]; + } + niter++; + // depsilon_dmu is now fully iterated. + + // Calculate the diff: + depsilon_dmu.diff = 0.0; + for (int i = 0; i < depsilon_dmu.Npts; ++i) + depsilon_dmu.diff += fabs(depsilon_dmu.value[i] - depsilon_dmu.prev_value[i]); + + StopTime = clock(); + CPU_ticks += StopTime - StartTime; + + //cout << "depsilon_dmu: niter = " << niter << "\tdiff = " << depsilon_dmu.diff << endl; + } while (niter < 5 || niter < niter_max && CPU_ticks < Max_CPU_ticks && depsilon_dmu.diff > req_diff); + + return(depsilon_dmu); + } + + Root_Density LiebLin_rho_TBA (DP kBT, const Root_Density& epsilon, const Root_Density& depsilon_dmu) + { + Root_Density rho = epsilon; + for (int i = 0; i < epsilon.Npts; ++i) + rho.value[i] = -(1.0/twoPI) * depsilon_dmu.value[i] + * (epsilon.value[i] > 0.0 ? + exp(-epsilon.value[i]/kBT)/(1.0 + exp(-epsilon.value[i]/kBT)) : + 1.0/(1.0 + exp(epsilon.value[i]/kBT))); + + return(rho); + } + + Root_Density LiebLin_rhoh_TBA (DP kBT, const Root_Density& epsilon, const Root_Density& depsilon_dmu) + { + Root_Density rhoh = epsilon; + for (int i = 0; i < epsilon.Npts; ++i) + rhoh.value[i] = -(1.0/twoPI) * depsilon_dmu.value[i] + * (epsilon.value[i] > 0.0 ? + 1.0/(1.0 + exp(-epsilon.value[i]/kBT)) : + exp(epsilon.value[i]/kBT)/(1.0 + exp(epsilon.value[i]/kBT))); + + return(rhoh); + } + + DP LiebLin_nbar_TBA (const Root_Density& rho) + { + DP nbar = 0.0; + for (int i = 0; i < rho.Npts; ++i) nbar += rho.value[i] * rho.dlambda[i]; + + return(nbar); + } + + DP LiebLin_ebar_TBA (const Root_Density& rho) + { + DP ebar = 0.0; + for (int i = 0; i < rho.Npts; ++i) ebar += rho.lambda[i] * rho.lambda[i] * rho.value[i] * rho.dlambda[i]; + + return(ebar); + } + + DP LiebLin_sbar_TBA (const Root_Density& rho, const Root_Density& rhoh) + { + DP sbar = 0.0; + for (int i = 0; i < rho.Npts; ++i) sbar += ((rho.value[i] + rhoh.value[i]) * log(rho.value[i] + rhoh.value[i]) - rho.value[i] * log(rho.value[i]+1.0e-30) - rhoh.value[i] * log(rhoh.value[i]+1.0e-30)) * rho.dlambda[i]; + + return(sbar); + } + + LiebLin_Bethe_State Discretized_LiebLin_Bethe_State (DP c_int, DP L, int N, const Root_Density& rho) + { + // This function returns the Bethe state at finite size which is + // the closest approximation to the continuum density rho(lambda) + + // Each time N \int_{-\infty}^\lambda d\lambda' \rho(\lambda') crosses a half integer, add a particle: + DP integral = 0.0; + DP integral_prev = 0.0; + int Nfound = 0; + Vect lambda_found(0.0, 2*N); + + for (int i = 0; i < rho.Npts; ++i) { + integral_prev = integral; + integral += L * rho.value[i] * rho.dlambda[i]; + if (integral > Nfound + 0.5) { + // Subtle error: if the rho is too discontinuous, i.e. if more than one rapidity is found, must correct for this. + if (integral > Nfound + 1.5 && integral < Nfound + 2.5) { // found two rapidities + lambda_found[Nfound++] = 0.25 * (3.0 * rho.lambda[i-1] + rho.lambda[i]); + lambda_found[Nfound++] = 0.25 * (rho.lambda[i-1] + 3.0 * rho.lambda[i]); + } + else { + //lambda_found[Nfound] = rho.lambda[i]; + // Better: center the lambda_found between these points: + //lambda_found[Nfound] = rho.lambda[i-1] + (rho.lambda[i] - rho.lambda[i-1]) * ((Nfound + 1.0) - integral_prev)/(integral - integral_prev); + lambda_found[Nfound] = 0.5 * (rho.lambda[i-1] + rho.lambda[i]); + Nfound++; + } + } + //cout << "\ti = " << i << "\tintegral = " << integral << "\tNfound = " << Nfound << endl; + } + //cout << "rho: " << rho.Npts << " points" << endl << rho.value << endl; + //cout << "sym: " << rho.value[0] << " " << rho.value[rho.value.size() - 1] + // << "\t" << rho.value[rho.value.size()/2] << " " << rho.value[rho.value.size()/2 + 1] << endl; + //cout << "Found " << Nfound << " particles." << endl; + //cout << "lambda_found = " << lambda_found << endl; + + Vect lambda(N); + // Fill up the found rapidities: + for (int il = 0; il < JSC::min(N, Nfound); ++il) lambda[il] = lambda_found[il]; + // If there are missing ones, put them at the end; ideally, this should never be called + for (int il = Nfound; il < N; ++il) lambda[il] = lambda_found[Nfound-1] + (il - Nfound + 1) * (lambda_found[Nfound-1] - lambda_found[Nfound-2]); + + // Calculate quantum numbers: 2\pi * (L lambda + \sum_j 2 atan((lambda - lambda_j)/c)) = I_j + // Use rounding. + Vect Ix2(N); + for (int i = 0; i < N; ++i) { + DP sum = 0.0; + for (int j = 0; j < N; ++j) sum += 2.0 * atan((lambda[i] - lambda[j])/c_int); + //Ix2[i] = 2.0* int((L * lambda[i] + sum)/twoPI) + (N % 2) - 1; + // For N is even/odd, we want to round off to the nearest odd/even integer. + Ix2[i] = 2.0 * floor((L* lambda[i] + sum)/twoPI + 0.5 * (N%2 ? 1 : 2)) + (N%2) - 1; + } + //cout << "Found quantum numbers " << endl << Ix2 << endl; + + // Check that the quantum numbers are all distinct: + bool allOK = false; + while (!allOK) { + for (int i = 0; i < N-1; ++i) if (Ix2[i] == Ix2[i+1] && Ix2[i] < 0) Ix2[i] -= 2; + for (int i = 1; i < N; ++i) if (Ix2[i] == Ix2[i-1] && Ix2[i] > 0) Ix2[i] += 2; + allOK = true; + for (int i = 0; i < N-1; ++i) if (Ix2[i] == Ix2[i+1]) allOK = false; + } + //cout << "Found modified quantum numbers " << endl << Ix2 << endl; + + LiebLin_Bethe_State rhostate(c_int, L, N); + rhostate.Ix2 = Ix2; + rhostate.Compute_All(true); + //cout << "rapidities of state found: " << rhostate.lambdaoc << endl; + + return(rhostate); + } + + +} // namespace JSC diff --git a/src/TBA/TBA_XXZ.cc b/src/TBA/TBA_XXZ.cc new file mode 100755 index 0000000..468c55e --- /dev/null +++ b/src/TBA/TBA_XXZ.cc @@ -0,0 +1,722 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c). + +----------------------------------------------------------- + +File: TBA_XXZ.cc + +Purpose: TBA for the gapless XXZ antiferromagnet. + + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + // First, define some useful kernels: + + DP XXZ_phi1_kernel (DP zeta, DP lambda) + { + return(2.0 * atan(tanh(lambda)/tan(0.5*zeta))); + } + + DP XXZ_phi2_kernel (DP zeta, DP lambda) + { + return(2.0 * atan(tanh(lambda)/tan(zeta))); + } + + DP XXZ_a1_kernel (DP sinzeta, DP coszeta, DP lambda) + { + DP expmin2abslambda = exp(-2.0 * fabs(lambda)); + + DP answer = sinzeta * expmin2abslambda + /(PI * (0.5 * (1.0 + expmin2abslambda * expmin2abslambda) - coszeta * expmin2abslambda)); + + return(answer); + } + + DP XXZ_da1dlambda_kernel (DP sinzeta, DP coszeta, DP lambda) + { + DP expmin2abslambda = exp(-2.0 * fabs(lambda)); + DP signlambda = lambda >= 0.0 ? 1.0 : -1.0; + DP answer = -sinzeta * signlambda * expmin2abslambda * (1.0 - expmin2abslambda * expmin2abslambda) + /(PI * pow((0.5 * (1.0 + expmin2abslambda * expmin2abslambda) + - coszeta * expmin2abslambda), 2.0)); + + return(answer); + } + + DP XXZ_a2_kernel (DP sin2zeta, DP cos2zeta, DP lambda) + { + DP expmin2abslambda = exp(-2.0 * fabs(lambda)); + + DP answer = sin2zeta * expmin2abslambda + /(PI * (0.5 * (1.0 + expmin2abslambda * expmin2abslambda) - cos2zeta * expmin2abslambda)); + + return(answer); + } + + void Iterate_XXZ_rhotot_GS (DP zeta, DP B, Root_Density& rhotot_GS) + { + // Performs an iteration of the TBA equation for rhotot_GS: + + // First, copy the actual values into the previous ones: + for (int i = 0; i < rhotot_GS.Npts; ++i) rhotot_GS.prev_value[i] = rhotot_GS.value[i]; + + // Calculate the new values: + DP conv = 0.0; + DP sinzeta = sin(zeta); + DP sin2zeta = sin(2.0*zeta); + DP coszeta = cos(zeta); + DP cos2zeta = cos(2.0*zeta); + for (int i = 0; i < rhotot_GS.Npts; ++i) { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < rhotot_GS.Npts; ++j) + conv += fabs(rhotot_GS.lambda[j]) > B ? 0.0 : + XXZ_a2_kernel (sin2zeta, cos2zeta, rhotot_GS.lambda[i] - rhotot_GS.lambda[j]) * rhotot_GS.prev_value[j] * rhotot_GS.dlambda[j]; + rhotot_GS.value[i] = XXZ_a1_kernel(sinzeta, coszeta, rhotot_GS.lambda[i]) - conv; + } + + // Calculate the sum of differences: + rhotot_GS.diff = 0.0; + for (int i = 0; i < rhotot_GS.Npts; ++i) + rhotot_GS.diff += fabs(rhotot_GS.value[i] - rhotot_GS.prev_value[i]); + + return; + } + + Root_Density XXZ_rhotot_GS (DP Delta, DP B, DP lambdamax, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the ground state + // root distribution of the XXZ model. + + if (Delta >= 1.0 || Delta <= 0.0) JSCerror("Delta out of bounds in XXZ_rhotot_GS."); + + DP zeta = acos(Delta); + + Root_Density rhotot_GS(Npts, lambdamax); + + // We now attempt to find a solution... + + int niter = 0; + do { + Iterate_XXZ_rhotot_GS (zeta, B, rhotot_GS); + //cout << rhotot_GS.diff << endl; + niter ++; + } while (rhotot_GS.diff > req_prec); + + //cout << "rhotot: niter = " << niter << endl; + return(rhotot_GS); + } + + DP Return_GS_Sz_tot_value (DP B, Root_Density& rhotot_GS) + { + DP integral = 0.0; + for (int i = 0; i < rhotot_GS.Npts; ++i) + integral += fabs(rhotot_GS.lambda[i]) > B ? 0.0 : rhotot_GS.value[i] * rhotot_GS.dlambda[i]; + + return(0.5 - integral); + } + + void Iterate_XXZ_eps_GS (DP zeta, DP Hz, Root_Density& eps_GS) + { + // Performs an iteration of the TBA equation for eps_GS: + + // First, copy the actual values into the previous ones: + for (int i = 0; i < eps_GS.Npts; ++i) eps_GS.prev_value[i] = eps_GS.value[i]; + + // Calculate the new values: + DP conv = 0.0; + DP sinzeta = sin(zeta); + DP sin2zeta = sin(2.0*zeta); + DP coszeta = cos(zeta); + DP cos2zeta = cos(2.0*zeta); + for (int i = 0; i < eps_GS.Npts; ++i) { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < eps_GS.Npts; ++j) + conv += eps_GS.prev_value[j] > 0.0 ? 0.0 : + XXZ_a2_kernel (sin2zeta, cos2zeta, eps_GS.lambda[i] - eps_GS.lambda[j]) + * eps_GS.prev_value[j] * eps_GS.dlambda[j]; + eps_GS.value[i] = Hz - PI * sinzeta * XXZ_a1_kernel(sinzeta, coszeta, eps_GS.lambda[i]) - conv; + //cout << i << "\t" << eps_GS.lambda[i] << "\t" << eps_GS.value[i] << "\t" << conv << endl; + } + + // Calculate the sum of differences: + eps_GS.diff = 0.0; + for (int i = 0; i < eps_GS.Npts; ++i) + eps_GS.diff += fabs(eps_GS.value[i] - eps_GS.prev_value[i]); + + return; + } + + Root_Density XXZ_eps_GS (DP Delta, DP Hz, DP lambdamax, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the ground state + // epsilon function of the XXZ model. + + if (Delta >= 1.0 || Delta <= 0.0) JSCerror("Delta out of bounds in XXZ_eps_GS."); + + DP zeta = acos(Delta); + + Root_Density eps_GS(Npts, lambdamax); + + // We now attempt to find a solution... + + int niter = 0; + do { + Iterate_XXZ_eps_GS (zeta, Hz, eps_GS); + //cout << eps_GS.diff << endl; + niter++; + //cout << niter << "\t" << eps_GS.diff << endl; + } while (eps_GS.diff > req_prec); + + //cout << "eps: niter = " << niter << endl; + return(eps_GS); + } + + void Iterate_XXZ_depsdlambda_GS (DP zeta, DP B, Root_Density& depsdlambda_GS) + { + // Performs an iteration of the TBA equation for depsdlambda_GS: + + // First, copy the actual values into the previous ones: + for (int i = 0; i < depsdlambda_GS.Npts; ++i) depsdlambda_GS.prev_value[i] = depsdlambda_GS.value[i]; + + // Calculate the new values: + DP conv = 0.0; + DP sinzeta = sin(zeta); + DP sin2zeta = sin(2.0*zeta); + DP coszeta = cos(zeta); + DP cos2zeta = cos(2.0*zeta); + for (int i = 0; i < depsdlambda_GS.Npts; ++i) { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < depsdlambda_GS.Npts; ++j) + conv += fabs(depsdlambda_GS.lambda[j]) > B ? 0.0 : + XXZ_a2_kernel (sin2zeta, cos2zeta, depsdlambda_GS.lambda[i] - depsdlambda_GS.lambda[j]) + * depsdlambda_GS.prev_value[j] * depsdlambda_GS.dlambda[j]; + depsdlambda_GS.value[i] = -PI * sinzeta * XXZ_da1dlambda_kernel(sinzeta, coszeta, depsdlambda_GS.lambda[i]) - conv; + } + + // Calculate the sum of differences: + depsdlambda_GS.diff = 0.0; + for (int i = 0; i < depsdlambda_GS.Npts; ++i) + depsdlambda_GS.diff += fabs(depsdlambda_GS.value[i] - depsdlambda_GS.prev_value[i]); + + return; + } + + Root_Density XXZ_depsdlambda_GS (DP Delta, DP B, DP lambdamax, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the ground state + // (d/d\lambda) epsilon function of the XXZ model. + + if (Delta >= 1.0 || Delta <= 0.0) JSCerror("Delta out of bounds in XXZ_rho_GS."); + + DP zeta = acos(Delta); + + Root_Density depsdlambda_GS(Npts, lambdamax); + + // We now attempt to find a solution... + + int niter = 0; + do { + Iterate_XXZ_depsdlambda_GS (zeta, B, depsdlambda_GS); + //cout << depsdlambda_GS.diff << endl; + niter++; + } while (depsdlambda_GS.diff > req_prec); + + //cout << "depsdlambda: niter = " << niter << endl; + return(depsdlambda_GS); + } + + void Iterate_XXZ_b2BB_lambda_B (DP zeta, DP B, Root_Density& b2BB) + { + // Calculates the vector corresponding to b2BB (lambda,B) + + // First, copy the actual values into the previous ones: + for (int i = 0; i < b2BB.Npts; ++i) b2BB.prev_value[i] = b2BB.value[i]; + + // Calculate the new values: + DP conv = 0.0; + DP sin2zeta = sin(2.0*zeta); + DP cos2zeta = cos(2.0*zeta); + for (int i = 0; i < b2BB.Npts; ++i) { + if (fabs(b2BB.lambda[i]) > B) b2BB.value[i] = 0.0; + else { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < b2BB.Npts; ++j) + conv += fabs(b2BB.lambda[j]) > B ? 0.0 : + XXZ_a2_kernel (sin2zeta, cos2zeta, b2BB.lambda[i] - b2BB.lambda[j]) + * b2BB.prev_value[j] * b2BB.dlambda[j]; + b2BB.value[i] = -XXZ_a2_kernel(sin2zeta, cos2zeta, b2BB.lambda[i] - B) + - conv; + } + } + // Calculate the sum of differences: + b2BB.diff = 0.0; + for (int i = 0; i < b2BB.Npts; ++i) + b2BB.diff += fabs(b2BB.value[i] - b2BB.prev_value[i]); + + return; + } + + Root_Density XXZ_b2BB_lambda_B (DP Delta, DP B, DP lambdamax, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the inverse + // kernel b2BB (lambda, B) used in Kbar function. + + if (Delta >= 1.0 || Delta <= 0.0) JSCerror("Delta out of bounds in XXZ_Kbackflow_GS."); + + DP zeta = acos(Delta); + + Root_Density b2BB_lambda_B(Npts, lambdamax); + + // We now attempt to find a solution... + + int niter = 0; + do { + Iterate_XXZ_b2BB_lambda_B (zeta, B, b2BB_lambda_B); + //cout << eps_GS.diff << endl; + niter++; + } while (b2BB_lambda_B.diff > req_prec); + + //if (lambda_p + lambda_h + 2.0*B > 0.0) cout << "Nonzero backflow possible." << endl; + //cout << "Kbackflow: niter = " << niter << endl; + return(b2BB_lambda_B); + } + + void Iterate_XXZ_b2BB_lambda_lambdap (DP zeta, DP B, DP lambdap, Root_Density& b2BB) + { + // Calculates the vector corresponding to b2BB (lambda,lambdap) + + // First, copy the actual values into the previous ones: + for (int i = 0; i < b2BB.Npts; ++i) b2BB.prev_value[i] = b2BB.value[i]; + + // Calculate the new values: + DP conv = 0.0; + DP sin2zeta = sin(2.0*zeta); + DP cos2zeta = cos(2.0*zeta); + for (int i = 0; i < b2BB.Npts; ++i) { + if (fabs(b2BB.lambda[i]) > B) b2BB.value[i] = 0.0; + else { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < b2BB.Npts; ++j) + conv += fabs(b2BB.lambda[j]) > B ? 0.0 : + XXZ_a2_kernel (sin2zeta, cos2zeta, b2BB.lambda[i] - b2BB.lambda[j]) + * b2BB.prev_value[j] * b2BB.dlambda[j]; + b2BB.value[i] = -XXZ_a2_kernel(sin2zeta, cos2zeta, b2BB.lambda[i] - lambdap) + - conv; + } + } + // Calculate the sum of differences: + b2BB.diff = 0.0; + for (int i = 0; i < b2BB.Npts; ++i) + b2BB.diff += fabs(b2BB.value[i] - b2BB.prev_value[i]); + + return; + } + + Root_Density XXZ_b2BB_lambda_lambdap (DP Delta, DP B, DP lambdap, DP lambdamax, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the inverse + // kernel b2BB (lambda, lambdap) used in Kbar function. + + if (Delta >= 1.0 || Delta <= 0.0) JSCerror("Delta out of bounds in XXZ_Kbackflow_GS."); + + DP zeta = acos(Delta); + + Root_Density b2BB_lambda_lambdap(Npts, lambdamax); + + // We now attempt to find a solution... + + int niter = 0; + do { + Iterate_XXZ_b2BB_lambda_lambdap (zeta, B, lambdap, b2BB_lambda_lambdap); + //cout << eps_GS.diff << endl; + niter++; + } while (b2BB_lambda_lambdap.diff > req_prec); + + //if (lambda_p + lambda_h + 2.0*B > 0.0) cout << "Nonzero backflow possible." << endl; + //cout << "Kbackflow: niter = " << niter << endl; + return(b2BB_lambda_lambdap); + } + + void Iterate_XXZ_Kbackflow_GS (DP zeta, DP B, DP lambda_p, DP lambda_h, Root_Density& Kbackflow_GS) + { + // Performs an iteration of the TBA equation for Kbackflow_GS: + + // First, copy the actual values into the previous ones: + for (int i = 0; i < Kbackflow_GS.Npts; ++i) Kbackflow_GS.prev_value[i] = Kbackflow_GS.value[i]; + + // Calculate the new values: + DP conv = 0.0; + DP sin2zeta = sin(2.0*zeta); + DP cos2zeta = cos(2.0*zeta); + for (int i = 0; i < Kbackflow_GS.Npts; ++i) { + //if (Kbackflow_GS.lambda[i] < -B || Kbackflow_GS.lambda[i] > lambda_p + lambda_h + B) Kbackflow_GS.value[i] = 0.0; + if (false && fabs(Kbackflow_GS.lambda[i]) > B) Kbackflow_GS.value[i] = 0.0; + else { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < Kbackflow_GS.Npts; ++j) + conv += fabs(Kbackflow_GS.lambda[j]) > B ? 0.0 : + //(Kbackflow_GS.lambda[i] < -B || Kbackflow_GS.lambda[i] > lambda_p + lambda_h + B) ? 0.0 : + XXZ_a2_kernel (sin2zeta, cos2zeta, Kbackflow_GS.lambda[i] - Kbackflow_GS.lambda[j]) + * Kbackflow_GS.prev_value[j] * Kbackflow_GS.dlambda[j]; + Kbackflow_GS.value[i] = -XXZ_a2_kernel(sin2zeta, cos2zeta, Kbackflow_GS.lambda[i] - lambda_p) + + XXZ_a2_kernel(sin2zeta, cos2zeta, Kbackflow_GS.lambda[i] - lambda_h) - conv; + } + } + // Calculate the sum of differences: + Kbackflow_GS.diff = 0.0; + for (int i = 0; i < Kbackflow_GS.Npts; ++i) + Kbackflow_GS.diff += fabs(Kbackflow_GS.value[i] - Kbackflow_GS.prev_value[i]); + + return; + } + + Root_Density XXZ_Kbackflow_GS (DP Delta, DP B, DP lambdamax, DP lambda_p, DP lambda_h, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the ground state + // backflow function of the XXZ model. + + if (Delta >= 1.0 || Delta <= 0.0) JSCerror("Delta out of bounds in XXZ_Kbackflow_GS."); + + DP zeta = acos(Delta); + + Root_Density Kbackflow_GS(Npts, lambdamax); + + // We now attempt to find a solution... + + int niter = 0; + do { + Iterate_XXZ_Kbackflow_GS (zeta, B, lambda_p, lambda_h, Kbackflow_GS); + //cout << eps_GS.diff << endl; + niter++; + } while (Kbackflow_GS.diff > req_prec); + + //if (lambda_p + lambda_h + 2.0*B > 0.0) cout << "Nonzero backflow possible." << endl; + //cout << "Kbackflow: niter = " << niter << endl; + return(Kbackflow_GS); + } + + void Iterate_XXZ_Fbackflow_GS (DP zeta, DP B, DP lambda_p, DP lambda_h, Root_Density& Fbackflow_GS) + { + // Performs an iteration of the TBA equation for Fbackflow_GS: + + // First, copy the actual values into the previous ones: + for (int i = 0; i < Fbackflow_GS.Npts; ++i) Fbackflow_GS.prev_value[i] = Fbackflow_GS.value[i]; + + // Calculate the new values: + DP conv = 0.0; + DP sin2zeta = sin(2.0*zeta); + DP cos2zeta = cos(2.0*zeta); + for (int i = 0; i < Fbackflow_GS.Npts; ++i) { + if (fabs(Fbackflow_GS.lambda[i]) > B) Fbackflow_GS.value[i] = 0.0; + else { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < Fbackflow_GS.Npts; ++j) + conv += fabs(Fbackflow_GS.lambda[j]) > B ? 0.0 : + XXZ_a2_kernel (sin2zeta, cos2zeta, Fbackflow_GS.lambda[i] - Fbackflow_GS.lambda[j]) + * Fbackflow_GS.prev_value[j] * Fbackflow_GS.dlambda[j]; + Fbackflow_GS.value[i] = (-XXZ_phi2_kernel(zeta, Fbackflow_GS.lambda[i] - lambda_p) + + XXZ_phi2_kernel(zeta, Fbackflow_GS.lambda[i] - lambda_h))/twoPI - conv; + } + } + // Calculate the sum of differences: + Fbackflow_GS.diff = 0.0; + for (int i = 0; i < Fbackflow_GS.Npts; ++i) + Fbackflow_GS.diff += fabs(Fbackflow_GS.value[i] - Fbackflow_GS.prev_value[i]); + + return; + } + + Root_Density XXZ_Fbackflow_GS (DP Delta, DP B, DP lambdamax, DP lambda_p, DP lambda_h, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the ground state + // F backflow function of the XXZ model. + + if (Delta >= 1.0 || Delta <= 0.0) JSCerror("Delta out of bounds in XXZ_Fbackflow_GS."); + + DP zeta = acos(Delta); + + Root_Density Fbackflow_GS(Npts, lambdamax); + + // We now attempt to find a solution... + + int niter = 0; + do { + Iterate_XXZ_Fbackflow_GS (zeta, B, lambda_p, lambda_h, Fbackflow_GS); + //cout << eps_GS.diff << endl; + niter++; + } while (Fbackflow_GS.diff > req_prec); + + //if (lambda_p + lambda_h + 2.0*B > 0.0) cout << "Nonzero backflow possible." << endl; + //cout << "Fbackflow: niter = " << niter << endl; + return(Fbackflow_GS); + } + + void Iterate_XXZ_Z_GS (DP zeta, DP B, Root_Density& Z_GS) + { + // Performs an iteration of the TBA equation for the dressed charge: + + // First, copy the actual values into the previous ones: + for (int i = 0; i < Z_GS.Npts; ++i) Z_GS.prev_value[i] = Z_GS.value[i]; + + // Calculate the new values: + DP conv = 0.0; + DP sin2zeta = sin(2.0*zeta); + DP cos2zeta = cos(2.0*zeta); + for (int i = 0; i < Z_GS.Npts; ++i) { + // First, calculate the convolution + conv = 0.0; + for (int j = 0; j < Z_GS.Npts; ++j) + conv += fabs(Z_GS.lambda[j]) > B ? 0.0 : + XXZ_a2_kernel (sin2zeta, cos2zeta, Z_GS.lambda[i] - Z_GS.lambda[j]) + * Z_GS.prev_value[j] * Z_GS.dlambda[j]; + Z_GS.value[i] = 1.0 - conv; + } + + // Calculate the sum of differences: + Z_GS.diff = 0.0; + for (int i = 0; i < Z_GS.Npts; ++i) + Z_GS.diff += fabs(Z_GS.value[i] - Z_GS.prev_value[i]); + + return; + } + + Root_Density XXZ_Z_GS (DP Delta, DP B, DP lambdamax, int Npts, DP req_prec) + { + + // This function returns a Root_Density object corresponding to the ground state + // dressed charge function of the XXZ model. + + if (Delta >= 1.0 || Delta <= 0.0) JSCerror("Delta out of bounds in XXZ_Z_GS."); + + DP zeta = acos(Delta); + + Root_Density Z_GS(Npts, lambdamax); + + // We now attempt to find a solution... + + do { + Iterate_XXZ_Z_GS (zeta, B, Z_GS); + //cout << depsdlambda_GS.diff << endl; + } while (Z_GS.diff > req_prec); + + return(Z_GS); + } + + /* + void XXZ_Compare_Lattice_and_Continuum_Backflows_base_1010 (DP Delta, int N, int M, long long int id) + { + // Define the chain: J, Delta, h, Nsites + Heis_Chain chain(1.0, Delta, 0.0, N); + + // Define the base: chain, Mdown + Heis_Base gbase(chain, M); + + // Define the ground state + XXZ_Bethe_State gstate(chain, gbase); + + // Compute everything about the ground state + gstate.Compute_All(true); + + // Define the number of rapidities for an excited state + Vect_INT Nrapidities(0, chain.Nstrings); + Nrapidities[0] = M; + + Vect_INT Nexcitations(0, 2* chain.Nstrings + 2); + Nexcitations[0] = 0; + Nexcitations[1] = 1; + Nexcitations[2] = 0; + Nexcitations[3] = 1; + + // Define a base configuration for this set of rapidity numbers + Heis_Base ebase (chain, Nrapidities); + + // Define the excited state + XXZ_Bethe_State estate(chain, ebase); + + // Define an offset from a base and a number of holes + Ix2_Offsets offsets(ebase, Nexcitations); + + // Set the offset to the desired id + offsets.Set_to_id (id); + + // Set the offset data into the quantum numbers + estate.Set_Ix2_Offsets(offsets); + + // Compute everything about this eigenstate + estate.Compute_All(true); + + DP k_ext_N = estate.K - gstate.K; + DP omega_N = estate.E - gstate.E; + DP lambdap = estate.lambda[0][0]; + DP lambdah = 0.0; + for (int alpha = 1; alpha < estate.base[0] - 1; ++alpha) + if (estate.Ix2[0][alpha + 1] > estate.Ix2[0][alpha] + 2) { + lambdah = 0.5 * (estate.lambda[0][alpha] + estate.lambda[0][alpha + 1]); + } + + + // Now solve the thermodynamic equations: + DP req_prec = 1.0e-12; + DP Hz_N = H_vs_M (Delta, N, M); + cout << "Hz_N = " << Hz_N << endl; + DP lambdamax = 4.0 * fabs(gstate.lambda[0][0]); + int Npts = 2000; + Root_Density eps_GS = XXZ_eps_GS (Delta, Hz_N, lambdamax, Npts, req_prec); + + // We can read off the value of B from the obtained eps_GS function: + DP B_eps = 0.0; + int iB_eps = Npts/2; + for (int i = Npts/2; i < Npts - 1; ++i) + if (eps_GS.value[i] * eps_GS.value[i+1] < 0.0) { + B_eps = 0.5 * (eps_GS.lambda[i] + eps_GS.lambda[i+1]); + iB_eps = i; + } + if (B_eps == 0.0) { + cout << "Delta = " << Delta << "\tN = " << N << "\tM = " << M << "\tHz_N = " << Hz_N << "\tid = " << id << endl; + JSCerror("B not found."); + } + if (B_eps > 0.5 * lambdamax) { + cout << "Delta = " << Delta << "\tN = " << N << "\tM = " << M << "\tHz_N = " << Hz_N << "\tid = " << id << endl; + JSCerror("Use a higher value of lambdamax."); + } + //DP lambdamax = lambdamax_eps; + //DP lambdamax = 2.0 * fabs(gstate.lambda[0][0]); // window of definition of Kbackflow is covered + // Start by finding the appropriate B: + DP B = DP(M)/N; // initial guess; + DP B_old = B; + DP dB = 0.5 * B; // variation we allow; + DP m_TBA = 0.0; + //Vect rhotot_GS_iter(10);// = XXZ_rhotot_GS (Delta, B, lambdamax, Npts, req_prec); + Root_Density rhotot_GS = XXZ_rhotot_GS (Delta, B, lambdamax, Npts, req_prec); + int iter = 0; + do { + B_old = B; + rhotot_GS = XXZ_rhotot_GS (Delta, B, lambdamax, Npts, req_prec); + m_TBA = 0.0; + for (int i1 = 0; i1 < rhotot_GS.Npts; ++i1) if (fabs(rhotot_GS.lambda[i1]) < B) m_TBA += rhotot_GS.value[i1] * rhotot_GS.dlambda[i1]; + if (m_TBA < DP(M)/N) B += dB; + else B -= dB; + cout << "B_old = " << B_old << "\tB = " << B << "\tdB = " << dB << "\tm = " << m_TBA << "\tM/N = " << DP(M)/N << endl; + dB *= 1.0/1.9; + } while (dB > 0.001 && iter < 9); + + cout << "Check of consistency: B_eps = " << B_eps << "\tB = " << B << "\tm_TBA = " << m_TBA << "\tM/N = " << DP(M)/N << endl; + + //Root_Density rhotot_GS = XXZ_rhotot_GS (Delta, B, lambdamax, Npts, req_prec); + Root_Density b2BB_lambda_B = XXZ_b2BB_lambda_B (Delta, B, lambdamax, Npts, req_prec); + Root_Density Kbackflow_GS = XXZ_Kbackflow_GS (Delta, B, lambdamax, lambdap, lambdah, Npts, req_prec); + Root_Density Fbackflow_GS = XXZ_Fbackflow_GS (Delta, B, lambdamax, lambdap, lambdah, Npts, req_prec); + + // Calculate the momentum and energy of the TBA state: + // Momentum: + DP conv = 0.0; + DP zeta = acos(Delta); + for (int l = 0; l < rhotot_GS.Npts; ++l) { + conv += fabs(rhotot_GS.lambda[l]) > B ? 0.0 : + (XXZ_phi2_kernel(zeta, lambdap - rhotot_GS.lambda[l]) + - XXZ_phi2_kernel(zeta, lambdah - rhotot_GS.lambda[l])) + * rhotot_GS.value[l] * rhotot_GS.dlambda[l]; + } + DP k_ext_TBA = -XXZ_phi1_kernel (zeta, lambdap) + XXZ_phi1_kernel (zeta, lambdah) + conv; + // Energy: + conv = 0.0; + DP sinzeta = sin(zeta); + DP coszeta = cos(zeta); + for (int l = 0; l < Kbackflow_GS.Npts; ++l) + conv += fabs(Kbackflow_GS.lambda[l]) > B ? 0.0 : + (Hz_N - PI * sinzeta * XXZ_a1_kernel(sinzeta, coszeta, Kbackflow_GS.lambda[l])) + * Kbackflow_GS.value[l] * Kbackflow_GS.dlambda[l]; + DP omega_TBA = -PI * sin(zeta) * (XXZ_a1_kernel(sinzeta, coszeta, lambdap) + - XXZ_a1_kernel(sinzeta, coszeta, lambdah)) + conv; + + cout << "k_ext_N = " << k_ext_N << "\tk_ext_TBA = " << k_ext_TBA << "\tomega_N = " << omega_N << "\tomega_TBA = " << omega_TBA << endl; + + // Get back to the finite lattice case: + // Define densities as 1/N \sum_j delta (lambda - lambda_j); use Gaussian smoothing of functions. + DP gwidth = 2.0 * B * log(DP(N))/M; // mean rapidity spacing times some factor, for smoothing of Gaussians + DP gwidthsq = gwidth * gwidth; + Vect_DP rhoGS_N (0.0, Npts); + Vect_DP rhoexc_N (0.0, Npts); + Vect_DP Kflow_N (0.0, Npts); + for (int iout = 0; iout < Npts; ++iout) { + DP lambdaout = rhotot_GS.lambda[iout]; + // We compute the densities at each point... + for (int alpha = 0; alpha < M; ++alpha) + rhoGS_N[iout] += exp(-(gstate.lambda[0][alpha] - lambdaout) * (gstate.lambda[0][alpha] - lambdaout)/gwidthsq); + rhoGS_N[iout] /= N * sqrt(PI) * gwidth; + for (int alpha = 1; alpha < estate.base[0]; ++alpha) + rhoexc_N[iout] += exp(-(estate.lambda[0][alpha] - lambdaout) * (estate.lambda[0][alpha] - lambdaout)/gwidthsq); + rhoexc_N[iout] /= N * sqrt(PI) * gwidth; + + Kflow_N[iout] = N * (rhoexc_N[iout] - rhoGS_N[iout] + (1.0/N) * exp(-(lambdah - lambdaout) * (lambdah - lambdaout)/gwidthsq)/(sqrt(PI) * gwidth)); + } // for iout + + cout << "Here 1" << endl; + // Now produce a file with the density flow: + stringstream flo_stringstream; + flo_stringstream << "Flow_D_" << Delta << "_N_" << N << "_M_" << M << "_base_id_" << estate.base_id << "_type_id_" << estate.type_id + << "_id_" << estate.id << ".flo"; + string flo_string = flo_stringstream.str(); + const char* flo_Cstr = flo_string.c_str(); + + ofstream outfile_flo; + outfile_flo.open(flo_Cstr); + outfile_flo.precision(16); + + for (int iout = 0; iout < Npts; ++iout) + outfile_flo << rhotot_GS.lambda[iout] << "\t" << rhotot_GS.value[iout] << "\t" << rhoGS_N[iout] << "\t" + << rhoexc_N[iout] << "\t" << Kbackflow_GS.value[iout] << "\t" + << (fabs(Kbackflow_GS.lambda[iout]) > B ? 0.0 : Kbackflow_GS.value[iout]) - b2BB_lambda_B.value[iout] * Fbackflow_GS.value[iB_eps] + + b2BB_lambda_B.value[b2BB_lambda_B.Npts - 1 - iout] * Fbackflow_GS.value[Fbackflow_GS.Npts - 1 - iB_eps] << "\t" << Kflow_N[iout] << endl; + + outfile_flo.close(); + + // ... and a file with the remainder of the data: + stringstream dat_stringstream; + dat_stringstream << "Flow_D_" << Delta << "_N_" << N << "_M_" << M << "_base_id_" << estate.base_id << "_type_id_" << estate.type_id + << "_id_" << estate.id << ".dat"; + string dat_string = dat_stringstream.str(); + const char* dat_Cstr = dat_string.c_str(); + + ofstream outfile_dat; + outfile_dat.open(dat_Cstr); + outfile_dat.precision(16); + + outfile_dat << "lambdap\tlambdah\tk_ext_N\tk_ext_TBA\tomega_N\tomega_TBA" + << lambdap << "\t" << lambdah << "\t" << k_ext_N << "\t" << k_ext_TBA << "\t" << omega_N << "\t" << omega_TBA << endl; + + outfile_dat.close(); + + return; + } + */ + +} // namespace JSC + + + diff --git a/src/UTILS/Data_File_Name.cc b/src/UTILS/Data_File_Name.cc new file mode 100644 index 0000000..67dfba3 --- /dev/null +++ b/src/UTILS/Data_File_Name.cc @@ -0,0 +1,210 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/SCAN/Data_File_Name.cc + +Purpose: defines functions returning stringstreams for + standardised file names. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + // For each model, two functions are given: the first uses physical parameters as arguments, + // and can be called with functions `knowing' which model is dealt with. + // The second is overloaded using the Bethe_State objects, and can be called in functions + // which are universal to all models. + + + // Lieb-Liniger: + + void Data_File_Name (stringstream& name, char whichDSF, DP c_int, DP L, int N, int iKmin, int iKmax, DP kBT, DP L2, string defaultScanStatename) + { + name << "LiebLin_"; + if (whichDSF == 'Z') name << "Z"; + else if (whichDSF == 'd') name << "Rho_Rho"; + else if (whichDSF == 'g') name << "Psi_Psidag"; + else if (whichDSF == 'o') name << "Psidag_Psi"; + else if (whichDSF == 'q') name << "GeomQuench"; + else if (whichDSF == '1') name << "Type_I_Exp_Data"; + else if (whichDSF == 'B') name << "BECg2"; + else if (whichDSF == 'C') name << "BECoverlap"; + else JSCerror("Option not implemented in Data_File_Name"); + + name << "_c_" << c_int << "_L_" << L << "_N_" << N; + //if (fixed_iK) name << "_iK_" << iKneeded; + if (defaultScanStatename == "") name << "_" << N << "_0_"; // simulates label of ground state + else name << "_" << defaultScanStatename; + if (iKmin == iKmax) name << "_iK_" << iKmin; + else name << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + if (kBT > 0.0) name << "_kBT_" << kBT; + if (whichDSF == 'q') name << "_L2_" << L2; + + return; + } + + //void Data_File_Name (stringstream& name, char whichDSF, bool fixed_iK, int iKneeded, LiebLin_Bethe_State& State, LiebLin_Bethe_State& RefScanState) + void Data_File_Name (stringstream& name, char whichDSF, int iKmin, int iKmax, DP kBT, LiebLin_Bethe_State& State, LiebLin_Bethe_State& RefScanState, string defaultScanStatename) + { + name << "LiebLin_"; + if (whichDSF == 'Z') name << "Z"; + else if (whichDSF == 'd') name << "Rho_Rho"; + else if (whichDSF == 'g') name << "Psi_Psidag"; + else if (whichDSF == 'o') name << "Psidag_Psi"; + else if (whichDSF == 'q') name << "GeomQuench"; + else if (whichDSF == '1') name << "Type_I_Exp_Data"; + else if (whichDSF == 'B') name << "BECg2"; + else if (whichDSF == 'C') name << "BECoverlap"; + else JSCerror("Option not implemented in Data_File_Name"); + + name << "_c_" << State.c_int << "_L_" << State.L << "_N_" << State.N; + //if (fixed_iK) name << "_iK_" << iKneeded; + if (defaultScanStatename == "") name << "_" << State.label; + else name << "_" << defaultScanStatename; + if (iKmin == iKmax) name << "_iK_" << iKmin; else name << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + if (kBT > 0.0) name << "_kBT_" << kBT; + if (whichDSF == 'q') name << "_L2_" << RefScanState.L; + + return; + } + + + // Heisenberg: + + //void Data_File_Name (stringstream& name, char whichDSF, DP Delta, int N, int M, bool fixed_iK, int iKneeded, int N2) + void Data_File_Name (stringstream& name, char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, DP kBT, int N2, string defaultScanStatename) + { + name << "HEIS_"; + if (whichDSF == 'Z') name << "Z"; + else if (whichDSF == 'm') name << "Smp"; + else if (whichDSF == 'z') name << "Szz"; + else if (whichDSF == 'p') name << "Spm"; + else if (whichDSF == 'a') name << "SzSz"; + else if (whichDSF == 'b') name << "SzSm"; + else if (whichDSF == 'c') name << "SmSm"; + else if (whichDSF == 'q') name << "GeomQuench"; + else { + cout << "Option tried (1): " << whichDSF << endl; + JSCerror("Option not implemented in Data_File_Name"); + } + + name << "_D_" << Delta << "_N_" << N << "_M_"; + for (int i = 0; i < int(log10(DP(N/2))) - int(log10(DP(M))); ++i) name << "0"; + name << M; + if (defaultScanStatename == "") name << "_" << M << "_0_"; // simulates label of ground state + else name << "_" << defaultScanStatename; + + //if (fixed_iK) { + // name << "_iK_"; + // for (int i = 0; i < int(log10(DP(N/2))) - int(log10(DP(iKneeded))); ++i) name << "0"; + // name << iKneeded; + //} + // From ABACUS++G_8 onwards: for Heisenberg, always scan over all momentum integers + //if (iKmin == iKmax) name << "_iK_" << iKmin; else name << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + if (kBT > 0.0) name << "_kBT_" << kBT; + if (whichDSF == 'q') name << "_N2_" << N2; + + return; + } + + //void Data_File_Name (stringstream& name, char whichDSF, bool fixed_iK, int iKneeded, Heis_Bethe_State& State, Heis_Bethe_State& RefScanState) + void Data_File_Name (stringstream& name, char whichDSF, int iKmin, int iKmax, DP kBT, Heis_Bethe_State& State, Heis_Bethe_State& RefScanState, string defaultScanStatename) + { + name << "HEIS_"; + if (whichDSF == 'Z') name << "Z"; + else if (whichDSF == 'm') name << "Smp"; + else if (whichDSF == 'z') name << "Szz"; + else if (whichDSF == 'p') name << "Spm"; + else if (whichDSF == 'a') name << "SzSz"; + else if (whichDSF == 'b') name << "SzSm"; + else if (whichDSF == 'c') name << "SmSm"; + else if (whichDSF == 'q') name << "GeomQuench"; + else { + cout << "Option tried (2): " << whichDSF << endl; + JSCerror("Option not implemented in Data_File_Name"); + } + + name << "_D_" << State.chain.Delta << "_N_" << State.chain.Nsites << "_M_"; + for (int i = 0; i < int(log10(DP(State.chain.Nsites/2))) - int(log10(DP(State.base.Mdown))); ++i) name << "0"; + name << State.base.Mdown; + if (defaultScanStatename == "") name << "_" << State.label; + else name << "_" << defaultScanStatename; + /* + if (fixed_iK) { + name << "_iK_"; + for (int i = 0; i < int(log10(DP(State.chain.Nsites/2))) - int(log10(DP(iKneeded))); ++i) name << "0"; + name << iKneeded; + } + */ + // From ABACUS++G_8 onwards: for Heisenberg, always scan over all momentum integers + //if (iKmin == iKmax) name << "_iK_" << iKmin; else name << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + if (kBT > 0.0) name << "_kBT_" << kBT; + if (whichDSF == 'q') name << "_N2_" << RefScanState.chain.Nsites; + + return; + } + + + // One-D spinless fermions: + /* + void ODSLF_Data_File_Name (stringstream& name, char whichDSF, DP Delta, int N, int M, int iKmin, int iKmax, DP kBT, int N2, string defaultScanStatename) + { + name << "ODSLF_"; + if (whichDSF == 'Z') name << "Z"; + else if (whichDSF == 'm') name << "cdag_c"; + else if (whichDSF == 'z') name << "zz"; + else if (whichDSF == 'p') name << "c_cdag"; + else if (whichDSF == 'q') name << "GeomQuench"; + else JSCerror("Option not implemented in Data_File_Name"); + + name << "_D_" << Delta << "_N_" << N << "_M_"; + for (int i = 0; i < int(log10(DP(N/2))) - int(log10(DP(M))); ++i) name << "0"; + name << M; + + if (iKmin == iKmax) name << "_iK_" << iKmin; else name << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + if (kBT > 0.0) name << "_kBT_" << kBT; + if (whichDSF == 'q') name << "_N2_" << N2; + + return; + } + + //void Data_File_Name (stringstream& name, char whichDSF, bool fixed_iK, int iKneeded, Heis_Bethe_State& State, Heis_Bethe_State& RefScanState) + void Data_File_Name (stringstream& name, char whichDSF, int iKmin, int iKmax, DP kBT, ODSLF_Bethe_State& State, ODSLF_Bethe_State& RefScanState, string defaultScanStatename) + { + name << "ODSLF_"; + if (whichDSF == 'Z') name << "Z"; + else if (whichDSF == 'm') name << "cdag_c"; + else if (whichDSF == 'z') name << "zz"; + else if (whichDSF == 'p') name << "c_cdag"; + //else if (whichDSF == 'a') name << "SzSz"; + //else if (whichDSF == 'b') name << "SzSm"; + //else if (whichDSF == 'q') name << "GeomQuench"; + else JSCerror("Option not implemented in Data_File_Name"); + + name << "_D_" << State.chain.Delta << "_N_" << State.chain.Nsites << "_M_"; + for (int i = 0; i < int(log10(DP(State.chain.Nsites/2))) - int(log10(DP(State.base.Mdown))); ++i) name << "0"; + name << State.base.Mdown; + if (fixed_iK) { + name << "_iK_"; + for (int i = 0; i < int(log10(DP(State.chain.Nsites/2))) - int(log10(DP(iKneeded))); ++i) name << "0"; + name << iKneeded; + } + if (iKmin == iKmax) name << "_iK_" << iKmin; else name << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + if (kBT > 0.0) name << "_kBT_" << kBT; + if (whichDSF == 'q') name << "_N2_" << RefScanState.chain.Nsites; + + return; + } +*/ + +} // namespace JSC diff --git a/src/UTILS/K_and_Omega_Files.cc b/src/UTILS/K_and_Omega_Files.cc new file mode 100644 index 0000000..53f8fc1 --- /dev/null +++ b/src/UTILS/K_and_Omega_Files.cc @@ -0,0 +1,65 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: K_and_Omega_Files.cc + +Purpose: Utilities: momentum and frequency files. + +***********************************************************/ + +#include "JSC.h" + +namespace JSC { + + void Write_K_File (DP Length, int iKmin, int iKmax) + { + stringstream K_file; + string K_file_string; + K_file << "K_Length_" << Length << "_iKmin_" << iKmin << "_iKmax_" << iKmax << ".dat"; + K_file_string = K_file.str(); + const char* K_file_Cstr = K_file_string.c_str(); + + ofstream outfile_K; + + outfile_K.open(K_file_Cstr); + + outfile_K.setf(ios::fixed); + outfile_K.setf(ios::showpoint); + outfile_K.precision(16); + + for (int iK = iKmin; iK <= iKmax; ++iK) outfile_K << 2.0*PI * iK/Length << endl; + + outfile_K.close(); + + return; + } + + void Write_Omega_File (int Nout_omega, DP omegamin, DP omegamax) + { + stringstream w_file; + string w_file_string; + w_file << "Omega_ommin_" << omegamin << "_ommax_" << omegamax << "_Nom_" << Nout_omega << ".dat"; + w_file_string = w_file.str(); + const char* w_file_Cstr = w_file_string.c_str(); + + ofstream outfile_w; + + outfile_w.open(w_file_Cstr); + + outfile_w.setf(ios::fixed); + outfile_w.setf(ios::showpoint); + outfile_w.precision(16); + + for (int iw = 0; iw < Nout_omega; ++iw) outfile_w << omegamin + (iw + 0.5) * (omegamax - omegamin)/Nout_omega << endl; + + outfile_w.close(); + + return; + } + +} // namespace JSC diff --git a/src/UTILS/Smoothen_RAW_into_SF.cc b/src/UTILS/Smoothen_RAW_into_SF.cc new file mode 100644 index 0000000..3ac6d7d --- /dev/null +++ b/src/UTILS/Smoothen_RAW_into_SF.cc @@ -0,0 +1,419 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Smoothen_RAW_into_SF.cc + +Purpose: from a .raw file, produces .dsf (dynamical sf) and .ssf (static sf) files. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + DP Smoothen_RAW_into_SF (string prefix, int iKmin, int iKmax, int DiK, + DP ommin, DP ommax, int Nom, DP gwidth, DP normalization, DP denom_sum_K) + { + // ommax is omega max for .dsf file, Nom is the number of omega slots used. + + // gwidth is the width of the smoothing Gaussian, defined as + // exp(-omega^2/(2 * gwidth^2)) + + // DiK is the (half-)window in iK which is averaged over. Averaging over a single iK means DiK == 0. + + // Open the original raw file: + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + ifstream RAW_infile; + RAW_infile.open(RAW_Cstr); + if (RAW_infile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("Could not open RAW_infile... "); + } + + if (iKmax - iKmin + 1 < 0) JSCerror("Improper iKmin, iKmax in Smoothen_RAW_into_SF"); + RecMat DSF(Nom, iKmax - iKmin + 1); + Vect_DP SSF(0.0, iKmax - iKmin + 1); + Vect_DP ASF(0.0, Nom); + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + // Momenta: average over 2*DiK + 1 entries. Weigh them linearly decreasing away from central one. + // Setting central one to value 1 + DiK, + // total weight is 1 + DiK + 2* \sum_1^DiK n = 1 + DiK + DiK (DiK + 1) = (DiK + 1)^2. + // Weight given is thus abs(DiK + 1 - (iK - iK'))/(DiK + 1)^2 for abs(iK - iK') <= DiK. + Vect_DP Kweight(DiK + 1); + for (int i = 0; i < DiK + 1; ++i) Kweight[i] = (DiK + 1.0 - i)/pow(DiK + 1.0, 2); + + Vect_DP omegaout (Nom); + for (int i = 0; i < Nom; ++i) omegaout[i] = ommin + (0.5 + i) * (ommax - ommin)/Nom; + + DP d_omega; + DP big_gwidth_used = 10.0 * gwidth; // neglect terms having gaussian < exp(-50) + DP oneovertwowidthsq = 1.0/(2.0 * gwidth * gwidth); + DP SFfactor = 1.0; + + while (RAW_infile.peek() != EOF) { + //RAW_infile >> omega >> iK >> FF >> conv >> label; + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (iK >= iKmin && iK <= iKmax && fabs(omega) > 1.0e-8) { // remove connected part of DSF + //SSF[iK - iKmin] += FF * FF; + for (int deltaiK = -DiK; deltaiK <= DiK; ++deltaiK) + if (iK + deltaiK >= iKmin && iK + deltaiK <= iKmax) + SSF[iK + deltaiK - iKmin] += Kweight[abs(deltaiK)] * FF * FF; + for (int iomega = 0; iomega < Nom; ++iomega) + if (big_gwidth_used > (d_omega = fabs(omegaout[iomega] - omega))) { + //DSF[iomega][iK - iKmin] += FF * FF * exp(-d_omega*d_omega * oneovertwowidthsq); + SFfactor = FF * FF * exp(-d_omega*d_omega * oneovertwowidthsq); + ASF[iomega] += SFfactor; + if (fabs(omega) > 1.0e-12) // exclude the delta function contribution coming from diagonal term, if present + for (int deltaiK = -DiK; deltaiK <= DiK; ++deltaiK) + if (iK + deltaiK >= iKmin && iK + deltaiK <= iKmax) + DSF[iomega][iK + deltaiK - iKmin] += Kweight[abs(deltaiK)] * SFfactor; + } + } + } + RAW_infile.close(); + + // Reset proper normalization: + DP normalization_used = normalization * 1.0/(sqrt(twoPI) * gwidth); // Gaussian factor + + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) { + SSF[iK] *= normalization/twoPI; // twoPI from integral over omega: \int d\omega/2\pi + for (int iomega = 0; iomega < Nom; ++iomega) + DSF[iomega][iK] *= normalization_used; + } + + DP ASFnormalization = normalization_used/denom_sum_K; + for (int iomega = 0; iomega < Nom; ++iomega) ASF[iomega] *= ASFnormalization; + + + // Output to .dsf, .ssf and .asf files + + stringstream DSF_stringstream; string DSF_string; + DSF_stringstream << prefix; + //if (iKmax != iKmin) DSF_stringstream << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + if (DiK > 0) DSF_stringstream << "_DiK_" << DiK; + DSF_stringstream << "_ommin_"<< ommin << "_ommax_" << ommax << "_Nom_" << Nom << "_w_" << gwidth << ".dsf"; + DSF_string = DSF_stringstream.str(); const char* DSF_Cstr = DSF_string.c_str(); + + ofstream DSF_outfile; + DSF_outfile.open(DSF_Cstr); + DSF_outfile.precision(12); + + for (int iomega = 0; iomega < Nom; ++iomega) { + if (iomega > 0) DSF_outfile << endl; + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) + DSF_outfile << DSF[iomega][iK] << "\t"; + } + DSF_outfile.close(); + + stringstream SSF_stringstream; string SSF_string; + SSF_stringstream << prefix; + //if (iKmin != iKmax) SSF_stringstream << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + if (DiK > 0) SSF_stringstream << "_DiK_" << DiK; + SSF_stringstream << ".ssf"; + SSF_string = SSF_stringstream.str(); const char* SSF_Cstr = SSF_string.c_str(); + + ofstream SSF_outfile; + SSF_outfile.open(SSF_Cstr); + SSF_outfile.precision(12); + + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) { + if (iK > 0) SSF_outfile << endl; + SSF_outfile << iK + iKmin << "\t" << SSF[iK]; + } + SSF_outfile.close(); + + + stringstream ASF_stringstream; string ASF_string; + ASF_stringstream << prefix; + //if (iKmax != iKmin) DSF_stringstream << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + ASF_stringstream << "_ommin_"<< ommin << "_ommax_" << ommax << "_Nom_" << Nom << "_w_" << gwidth << ".asf"; + ASF_string = ASF_stringstream.str(); const char* ASF_Cstr = ASF_string.c_str(); + + ofstream ASF_outfile; + ASF_outfile.open(ASF_Cstr); + ASF_outfile.precision(12); + + for (int iomega = 0; iomega < Nom; ++iomega) { + if (iomega > 0) ASF_outfile << endl; + ASF_outfile << omegaout[iomega] << "\t" << ASF[iomega]; + } + ASF_outfile.close(); + + + + // Check sums: + DP sumdsf = 0.0; + DP sumssf = 0.0; + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) { + sumssf += SSF[iK]; + for (int iomega = 0; iomega < Nom; ++iomega) + sumdsf += DSF[iomega][iK]; + } + sumssf /= (iKmax - iKmin + 1); + sumdsf /= (iKmax - iKmin + 1) * Nom; + + return(sumdsf); + } + + // This is the same function as above, but now using data for a diagonal ensemble of states + DP Smoothen_RAW_into_SF (string prefix, Vect rawfilename, Vect weight, int iKmin, int iKmax, int DiK, + DP ommin, DP ommax, int Nom, DP gwidth, DP normalization, DP denom_sum_K) + { + // ommax is omega max for .dsf file, Nom is the number of omega slots used. + + // gwidth is the width of the smoothing Gaussian, defined as + // exp(-omega^2/(2 * gwidth^2)) + + // DiK is the (half-)window in iK which is averaged over. Averaging over a single iK means DiK == 0. + + if (iKmax - iKmin + 1 < 0) JSCerror("Improper iKmin, iKmax in Smoothen_RAW_into_SF"); + RecMat DSF(Nom, iKmax - iKmin + 1); + Vect_DP SSF(0.0, iKmax - iKmin + 1); + Vect_DP ASF(0.0, Nom); + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + // Momenta: average over 2*DiK + 1 entries. Weigh them linearly decreasing away from central one. + // Setting central one to value 1 + DiK, + // total weight is 1 + DiK + 2* \sum_1^DiK n = 1 + DiK + DiK (DiK + 1) = (DiK + 1)^2. + // Weight given is thus abs(DiK + 1 - (iK - iK'))/(DiK + 1)^2 for abs(iK - iK') <= DiK. + Vect_DP Kweight(DiK + 1); + for (int i = 0; i < DiK + 1; ++i) Kweight[i] = (DiK + 1.0 - i)/pow(DiK + 1.0, 2); + + Vect_DP omegaout (Nom); + for (int i = 0; i < Nom; ++i) omegaout[i] = ommin + (0.5 + i) * (ommax - ommin)/Nom; + + DP d_omega; + DP big_gwidth_used = 10.0 * gwidth; // neglect terms having gaussian < exp(-50) + DP oneovertwowidthsq = 1.0/(2.0 * gwidth * gwidth); + DP SFfactor = 1.0; + + for (int ns = 0; ns < weight.size(); ++ns) { + + // Open the original raw file: + //stringstream RAW_stringstream; string RAW_string; + //RAW_stringstream << prefix << ".raw"; + //RAW_string = RAW_stringstream.str(); + //const char* RAW_Cstr = RAW_string.c_str(); + const char* RAW_Cstr = rawfilename[ns].c_str(); + + ifstream RAW_infile; + RAW_infile.open(RAW_Cstr); + if (RAW_infile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("Could not open RAW_infile... "); + } + + while (RAW_infile.peek() != EOF) { + //RAW_infile >> omega >> iK >> FF >> conv >> label; + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (iK >= iKmin && iK <= iKmax && fabs(omega) > 1.0e-8) { // remove connected part of DSF) + SSF[iK - iKmin] += weight[ns] * FF * FF; + for (int iomega = 0; iomega < Nom; ++iomega) + if (big_gwidth_used > (d_omega = fabs(omegaout[iomega] - omega))) { + //DSF[iomega][iK - iKmin] += FF * FF * exp(-d_omega*d_omega * oneovertwowidthsq); + SFfactor = weight[ns] * FF * FF * exp(-d_omega*d_omega * oneovertwowidthsq); + ASF[iomega] += SFfactor; + if (fabs(omega) > 1.0e-12) // exclude the delta function contribution coming from diagonal term, if present + for (int deltaiK = -DiK; deltaiK <= DiK; ++deltaiK) + if (iK + deltaiK >= iKmin && iK + deltaiK <= iKmax) + DSF[iomega][iK + deltaiK - iKmin] += Kweight[abs(deltaiK)] * SFfactor; + } + } + } + RAW_infile.close(); + } // for ns + + // Reset proper normalization: + DP normalization_used = normalization * 1.0/(sqrt(twoPI) * gwidth); // Gaussian factor + + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) { + SSF[iK] *= normalization/twoPI; // twoPI from integral over omega: \int d\omega/2\pi + for (int iomega = 0; iomega < Nom; ++iomega) + DSF[iomega][iK] *= normalization_used; + } + + DP ASFnormalization = normalization_used/denom_sum_K; + for (int iomega = 0; iomega < Nom; ++iomega) ASF[iomega] *= ASFnormalization; + + + // Output to .dsf, .ssf and .asf files + + stringstream DSF_stringstream; string DSF_string; + DSF_stringstream << prefix; + //if (iKmax != iKmin) DSF_stringstream << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + DSF_stringstream << "_ns_" << weight.size(); + if (DiK > 0) DSF_stringstream << "_DiK_" << DiK; + DSF_stringstream << "_ommin_"<< ommin << "_ommax_" << ommax << "_Nom_" << Nom << "_w_" << gwidth << ".dsf"; + DSF_string = DSF_stringstream.str(); const char* DSF_Cstr = DSF_string.c_str(); + + ofstream DSF_outfile; + DSF_outfile.open(DSF_Cstr); + DSF_outfile.precision(12); + + for (int iomega = 0; iomega < Nom; ++iomega) { + if (iomega > 0) DSF_outfile << endl; + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) + DSF_outfile << DSF[iomega][iK] << "\t"; + } + DSF_outfile.close(); + + stringstream SSF_stringstream; string SSF_string; + SSF_stringstream << prefix; + //if (iKmin != iKmax) SSF_stringstream << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + SSF_stringstream << "_ns_" << weight.size(); + SSF_stringstream << ".ssf"; + SSF_string = SSF_stringstream.str(); const char* SSF_Cstr = SSF_string.c_str(); + + ofstream SSF_outfile; + SSF_outfile.open(SSF_Cstr); + SSF_outfile.precision(12); + + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) { + if (iK > 0) SSF_outfile << endl; + SSF_outfile << iK + iKmin << "\t" << SSF[iK]; + } + SSF_outfile.close(); + + + stringstream ASF_stringstream; string ASF_string; + ASF_stringstream << prefix; + //if (iKmax != iKmin) DSF_stringstream << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + ASF_stringstream << "_ns_" << weight.size(); + ASF_stringstream << "_ommin_"<< ommin << "_ommax_" << ommax << "_Nom_" << Nom << "_w_" << gwidth << ".asf"; + ASF_string = ASF_stringstream.str(); const char* ASF_Cstr = ASF_string.c_str(); + + ofstream ASF_outfile; + ASF_outfile.open(ASF_Cstr); + ASF_outfile.precision(12); + + for (int iomega = 0; iomega < Nom; ++iomega) { + if (iomega > 0) ASF_outfile << endl; + ASF_outfile << omegaout[iomega] << "\t" << ASF[iomega]; + } + ASF_outfile.close(); + + + + // Check sums: + DP sumdsf = 0.0; + DP sumssf = 0.0; + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) { + sumssf += SSF[iK]; + for (int iomega = 0; iomega < Nom; ++iomega) + sumdsf += DSF[iomega][iK]; + } + sumssf /= (iKmax - iKmin + 1); + sumdsf /= (iKmax - iKmin + 1) * Nom; + + return(sumdsf); + } + + + + + DP Smoothen_RAW_into_ASF (string prefix, int iKmin, int iKmax, DP ommin, DP ommax, int Nom, DP gwidth, + DP normalization, DP denom_sum_K) + { + // ommax is omega max for .asf file, Nom is the number of omega slots used. + + // gwidth is the width of the smoothing Gaussian, defined as + // exp(-omega^2/(2 * gwidth^2)) + + // Open the original raw file: + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + ifstream RAW_infile; + RAW_infile.open(RAW_Cstr); + if (RAW_infile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("Could not open RAW_infile... "); + } + + if (iKmax - iKmin + 1 < 0) JSCerror("Improper iKmin, iKmax in Smoothen_RAW_into_ASF"); + Vect_DP ASF(Nom); + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + Vect_DP omegaout (Nom); + for (int i = 0; i < Nom; ++i) omegaout[i] = ommin + (0.5 + i) * (ommax - ommin)/Nom; + + DP d_omega; + DP big_gwidth_used = 10.0 * gwidth; // neglect terms having gaussian < exp(-50) + DP oneovertwowidthsq = 1.0/(2.0 * gwidth * gwidth); + + while (RAW_infile.peek() != EOF) { + //RAW_infile >> omega >> iK >> FF >> conv >> label; + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (iK >= iKmin && iK <= iKmax && fabs(omega) > 1.0e-8) { // remove connected part of DSF) + for (int iomega = 0; iomega < Nom; ++iomega) + if (big_gwidth_used > (d_omega = fabs(omegaout[iomega] - omega))) + ASF[iomega] += FF * FF * exp(-d_omega*d_omega * oneovertwowidthsq); + } + } + RAW_infile.close(); + + // Reset proper normalization: + DP normalization_used = normalization * 1.0/(sqrt(twoPI) * gwidth); // Gaussian factor + normalization_used /= denom_sum_K; + for (int iomega = 0; iomega < Nom; ++iomega) + ASF[iomega] *= normalization_used; + + // Output to .asf file + + stringstream ASF_stringstream; string ASF_string; + ASF_stringstream << prefix; + //if (iKmax != iKmin) DSF_stringstream << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + ASF_stringstream << "_ommin_"<< ommin << "_ommax_" << ommax << "_Nom_" << Nom << "_w_" << gwidth << ".asf"; + ASF_string = ASF_stringstream.str(); const char* ASF_Cstr = ASF_string.c_str(); + + ofstream ASF_outfile; + ASF_outfile.open(ASF_Cstr); + ASF_outfile.precision(12); + + for (int iomega = 0; iomega < Nom; ++iomega) { + if (iomega > 0) ASF_outfile << endl; + ASF_outfile << ASF[iomega] << "\t"; + } + ASF_outfile.close(); + + DP sumasf = 0.0; + for (int iomega = 0; iomega < Nom; ++iomega) + sumasf += ASF[iomega]; + + sumasf /= Nom; + + return(sumasf); + } + +} // namespace JSC diff --git a/src/UTILS/Smoothen_RAW_into_SF_LiebLin_Scaled.cc b/src/UTILS/Smoothen_RAW_into_SF_LiebLin_Scaled.cc new file mode 100644 index 0000000..e3ea934 --- /dev/null +++ b/src/UTILS/Smoothen_RAW_into_SF_LiebLin_Scaled.cc @@ -0,0 +1,175 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: Smoothen_RAW_into_SF_LiebLin_Scaled.cc + +Purpose: from a .raw file, produces .dsf (dynamical sf) file + with gaussian width a function of momentum, and a .ssf (static sf) file. + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + DP Smoothen_RAW_into_SF_LiebLin_Scaled (string prefix, DP L, int N, int iKmin, int iKmax, int DiK, + DP ommin, DP ommax, int Nom, DP width, DP normalization) + { + // DiK is the (half-)window in iK which is averaged over. A single iK means DiK == 0. + + // ommax is omega max for .dsf file, Nom is the number of omega slots used. + + // width is measured in units of the level spacing (for each momentum independently). + + // gwidth is the width of the smoothing Gaussian, defined as + // exp(-omega^2/(2 * gwidth^2)) + + // Open the original raw file: + stringstream RAW_stringstream; string RAW_string; + RAW_stringstream << prefix << ".raw"; + RAW_string = RAW_stringstream.str(); const char* RAW_Cstr = RAW_string.c_str(); + + ifstream RAW_infile; + RAW_infile.open(RAW_Cstr); + if (RAW_infile.fail()) { + cout << RAW_Cstr << endl; + JSCerror("Could not open RAW_infile... "); + } + + if (iKmax - iKmin + 1 < 0) JSCerror("Improper iKmin, iKmax in Smoothen_RAW_into_SF_LiebLin_Scaled"); + RecMat DSFS(Nom, iKmax - iKmin + 1); + Vect_DP SSF(0.0, iKmax - iKmin + 1); + + DP omega; + int iK; + DP FF; + //int conv; + DP dev; + string label; + + // Momenta: average over 2*DiK + 1 entries. Weigh them linearly decreasing away from central one. + // Setting central one to value 1 + DiK, + // total weight is 1 + DiK + 2* \sum_1^DiK n = 1 + DiK + DiK (DiK + 1) = (DiK + 1)^2. + // Weight given is thus abs(DiK + 1 - (iK - iK'))/(DiK + 1)^2 for abs(iK - iK') <= DiK. + Vect_DP Kweight(DiK + 1); + for (int i = 0; i < DiK + 1; ++i) Kweight[i] = (DiK + 1.0 - i)/pow(DiK + 1.0, 2); + + Vect_DP omegaout (Nom); + for (int i = 0; i < Nom; ++i) omegaout[i] = ommin + (0.5 + i) * (ommax - ommin)/Nom; + + DP d_omega; + Vect_DP gwidth (iKmax - iKmin + 1); + + // At fixed momentum k, we take the bandwidth as bw = e_1 (k) - e_2 (k) calculated in TG limit, + // i.e. bw = k^2 + 2\pi \rho k - (-k^2 + 2\pi\rho k) = 2 k^2 for 0 < k < 2k_F = 2\pi \rho + // and bw = 2 * (2\pi \rho)^2 for 2k_F < k. + + // For iK <= N, there are iK states in bw. + // For iK > N, there are N states. + + for (iK = iKmin; iK <= iKmax; ++iK) + //gwidth[iK - iKmin] = width * 2.0 * JSC::min( pow(twoPI * JSC::max(abs(iK),1)/L, 2.0), pow(twoPI * N/L, 2.0))/JSC::min(N, JSC::max(abs(iK), 1)); + // Make sure the width does not become lower than the omegaout raster: + gwidth[iK - iKmin] = JSC::max(2.0 * (ommax - ommin)/Nom, width * 2.0 * JSC::min( pow(twoPI * JSC::max(abs(iK),1)/L, 2.0), pow(twoPI * N/L, 2.0))/JSC::min(N, JSC::max(abs(iK), 1))); + + + Vect_DP big_gwidth_used (iKmax - iKmin + 1); + for (iK = iKmin; iK <= iKmax; ++iK) + big_gwidth_used[iK - iKmin] = 10.0 * gwidth[iK - iKmin]; // neglect terms having gaussian < exp(-50) + Vect_DP oneovertwowidthsq (iKmax - iKmin + 1); + for (iK = iKmin; iK <= iKmax; ++iK) + oneovertwowidthsq[iK - iKmin] = 1.0/(2.0 * gwidth[iK - iKmin] * gwidth[iK - iKmin]); + + // Reset proper normalization: + Vect_DP normalization_used (iKmax - iKmin + 1); + for (iK = iKmin; iK <= iKmax; ++iK) + normalization_used[iK - iKmin] = normalization * 1.0/(sqrt(twoPI) * gwidth[iK - iKmin]); // Gaussian factor + + DP FFsq = 0.0; + while (RAW_infile.peek() != EOF) { + //RAW_infile >> omega >> iK >> FF >> conv >> label; + RAW_infile >> omega >> iK >> FF >> dev >> label; + if (iK >= iKmin && iK <= iKmax && fabs(omega) > 1.0e-8) { // remove connected part of DSF + FFsq = FF * FF; + SSF[iK - iKmin] += FFsq; + for (int iomega = 0; iomega < Nom; ++iomega) + if (big_gwidth_used[iK - iKmin] > (d_omega = fabs(omegaout[iomega] - omega))) + //DSF[iomega][iK - iKmin] += FF * FF * normalization_used[iK - iKmin] * exp(-d_omega*d_omega * oneovertwowidthsq[iK - iKmin]); + for (int deltaiK = -DiK; deltaiK <= DiK; ++deltaiK) + if (iK + deltaiK >= iKmin && iK + deltaiK <= iKmax) + DSFS[iomega][iK + deltaiK - iKmin] += Kweight[abs(deltaiK)] * FFsq * normalization_used[iK + deltaiK - iKmin] + * exp(-d_omega*d_omega * oneovertwowidthsq[iK + deltaiK - iKmin]); + } + } + RAW_infile.close(); + + // Reset proper normalization: + //DP normalization_used = normalization * 1.0/(sqrt(twoPI) * gwidth); // Gaussian factor + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) { + SSF[iK] *= normalization/twoPI; + //FSR[iK] *= normalization_used; + //for (int iomega = 0; iomega < Nom; ++iomega) + //DSF[iomega][iK] *= normalization_used; + } + + // Output to .dsfs and .ssf files + + stringstream DSFS_stringstream; string DSFS_string; + DSFS_stringstream << prefix; + //if (iKmax != iKmin) DSF_stringstream << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + if (DiK > 0) DSFS_stringstream << "_DiK_" << DiK; + DSFS_stringstream << "_ommin_"<< ommin << "_ommax_" << ommax << "_Nom_" << Nom << "_w_" << width << ".dsfs"; + DSFS_string = DSFS_stringstream.str(); const char* DSFS_Cstr = DSFS_string.c_str(); + + ofstream DSFS_outfile; + DSFS_outfile.open(DSFS_Cstr); + DSFS_outfile.precision(12); + + for (int iomega = 0; iomega < Nom; ++iomega) { + if (iomega > 0) DSFS_outfile << endl; + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) + DSFS_outfile << DSFS[iomega][iK] << "\t"; + } + DSFS_outfile.close(); + + + stringstream SSF_stringstream; string SSF_string; + SSF_stringstream << prefix; + //if (iKmin != iKmax) SSF_stringstream << "_iKmin_" << iKmin << "_iKmax_" << iKmax; + SSF_stringstream << ".ssf"; + SSF_string = SSF_stringstream.str(); const char* SSF_Cstr = SSF_string.c_str(); + + ofstream SSF_outfile; + SSF_outfile.open(SSF_Cstr); + SSF_outfile.precision(12); + + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) { + SSF_outfile << iK + iKmin << "\t" << SSF[iK] << endl; + } + SSF_outfile.close(); + + + // Check sums: + DP sumdsf = 0.0; + //DP sumssf = 0.0; + for (int iK = 0; iK < iKmax - iKmin + 1; ++iK) { + //sumssf += FSR[iK]; + for (int iomega = 0; iomega < Nom; ++iomega) + sumdsf += DSFS[iomega][iK]; + } + //sumssf /= (iKmax - iKmin + 1); + sumdsf /= (iKmax - iKmin + 1) * Nom; + + return(sumdsf); + + } + +} // namespace JSC diff --git a/src/UTILS/Sort_RAW_File.cc b/src/UTILS/Sort_RAW_File.cc new file mode 100644 index 0000000..297b79a --- /dev/null +++ b/src/UTILS/Sort_RAW_File.cc @@ -0,0 +1,130 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: Sort_RAW_File.cc + +Purpose: orders the file in form factor or energy. + + +***********************************************************/ + +#include "JSC.h" + +namespace JSC { + + + void Sort_RAW_File (const char ffsq_file[], char optionchar) + { + Sort_RAW_File (ffsq_file, optionchar, 'a'); // dummy whichDSF, doesn't matter + } + + + void Sort_RAW_File (const char ff_file[], char optionchar, char whichDSF) + { + // Sorts FF in decreasing order for 'f' option, or energies in increasing order for option 'e', and writes .dat_srt file + + if (!(optionchar == 'e' || optionchar == 'f')) JSCerror("Wrong option in Sort_FF"); + + // Check size of threads file: + struct stat statbuf; + + stat (ff_file, &statbuf); + int filesize = statbuf.st_size; + + // Determine the number of entries approximately + int entry_size = 2* sizeof(float) + 2*sizeof(int); + + //const int MAXDATA = 50000000; + const int MAXDATA = filesize/entry_size + 10; + + DP* omega = new DP[MAXDATA]; + int* iK = new int[MAXDATA]; + DP* ff = new DP[MAXDATA]; + DP* ff_im = new DP[MAXDATA]; + //int* conv = new int[MAXDATA]; + DP* dev = new DP[MAXDATA]; + string* label = new string[MAXDATA]; + + ifstream infile; + infile.open(ff_file); + + if (infile.fail()) JSCerror("The input file was not opened successfully in Sort_RAW_File. "); + + stringstream outfilename; + string outfilename_string; + outfilename << ff_file << "_srt_" << optionchar; + outfilename_string = outfilename.str(); + const char* outfilename_c_str = outfilename_string.c_str(); + + ofstream outfile; + outfile.open(outfilename_c_str); + outfile.precision(16); + + int Ndata = 0; + while (((infile.peek()) != EOF) && (Ndata < MAXDATA)) { + + infile >> omega[Ndata]; + infile >> iK[Ndata]; + if (whichDSF != 'Z') infile >> ff[Ndata]; + if (whichDSF == 'q') infile >> ff_im[Ndata]; // imaginary part of overlap, quench case + //infile >> conv[Ndata]; + infile >> dev[Ndata]; + infile >> label[Ndata]; + + Ndata++; + } + infile.close(); + + int* index = new int[Ndata]; + DP* ffsq = new DP[Ndata]; + + for (int i = 0; i < Ndata; ++i) index[i] = i; + for (int i = 0; i < Ndata; ++i) ffsq[i] = ff[i] * ff[i]; + + if (optionchar == 'f') QuickSort(ffsq, index, 0, Ndata - 1); + else if (optionchar == 'e') QuickSort(omega, index, 0, Ndata - 1); + + for (int i = 0; i < Ndata; i++) { + + if (i > 0) outfile << endl; + if (optionchar == 'f') { + + outfile << omega[index[Ndata - 1 - i] ] << "\t" << iK[index[Ndata - 1 - i] ]; + if (whichDSF != 'Z') outfile << "\t" << ff[index[Ndata - 1 - i] ]; + if (whichDSF == 'q') outfile << "\t" << ff_im[index[Ndata - 1 - i] ]; + outfile << "\t" //<< conv[index[Ndata - 1 - i] ] << "\t" + << dev[index[Ndata - 1 - i] ] << "\t" + << label[index[Ndata - 1 - i] ]; + } + else if (optionchar == 'e') { + outfile << omega[i] << "\t" << iK[index[i] ]; + if (whichDSF != 'Z') outfile << "\t" << ff[index[i] ]; + if (whichDSF == 'q') outfile << "\t" << ff_im[index[i] ]; + outfile << "\t" //<< conv[index[i] ] << "\t" + << dev[index[i] ] << "\t" + << label[index[i] ]; + } + } + + outfile.close(); + + delete[] omega; + delete[] iK; + delete[] ff; + delete[] ff_im; + //delete[] conv; + delete[] dev; + delete[] label; + + delete[] index; + delete[] ffsq; + + return; + } + +} // namespace JSC diff --git a/src/UTILS/State_Label.cc b/src/UTILS/State_Label.cc new file mode 100644 index 0000000..b9e68ab --- /dev/null +++ b/src/UTILS/State_Label.cc @@ -0,0 +1,658 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS library. + +Copyright (c). + +----------------------------------------------------------- + +File: src/UTILS/State_Label.cc + +Purpose: universal implementation of state labels for ABACUS++ + + +***********************************************************/ + +#include "JSC.h" + +using namespace std; +using namespace JSC; + +namespace JSC { + + // The label of a state is built as follows: + + // M0[|type1:M1|type2:M2...]_nexc0[|nexc1|nexc2...]_type0Ix2old@type0Ix2new[:...][|type1Ix2old@type1Ix2new...] + + // A label is always relative to another label reference state, in practise + // the seed state used in the scanning for correlations (or otherwise by default: the ground state). + + // The first part of the label (before the first _ ) labels the particle content (the "base"). + // The second part (between the two _ ) specifies the number of quantum numbers + // excited to a position different from the configuration of the label reference state. + // The third part gives, for each excitation, the moved quantum number, and + // (after the @) the quantum number it has been moved to. + + // The second part is redundant (it could be read off the third part), but is kept since it is human-readable. + + // Example: + + // 300|2:23|5:4_3|2|1_-21@-43:11@-21:299@341|0@-24:4@12|5@35 + + // labels a state with + // 300 particles of type 0, 23 of type 2 and 4 of type 5 + // 3 type 0 particles displaced (Ix2 = -21 displaced to Ix2 = -43, 11 to -21 and 299 at 341) + // 2 type 2 particles displaced (Ix2 = 0 to -24 and 4 to 12) + // 1 type 5 particle displaced (Ix2 = 5 displaced to Ix2 = 35). + + + + string Extract_Base_Label (string label) + { + string::size_type i1 = label.find(LABELSEP); + + //cout << "Extracting Base_Label from label " << label << endl; + string baselabel = label.substr(0, i1); + //cout << "Extracted Base_Label " << baselabel << endl; + + return(baselabel); + } + + string Extract_nexc_Label (string label) + { + string::size_type i1 = label.find(LABELSEP); + string::size_type i2 = label.rfind(LABELSEP); + + return(label.substr(i1+1, i2-i1-1)); + } + + // For compressed labels: conversions between integers and char/strings. + // This is done according to the following data (in JSC_Scan.h): look for JSCcoding. + + string Convert_POSINT_to_STR (int int_to_convert) + { + // Converts a positive integer into a string according to the coding defined by global constant array JSCcoding. + if (int_to_convert < 0) JSCerror("Trying to convert a negative integer to a string."); + + int remainder = int_to_convert; + stringstream result_strstrm; + do { + //cout << "remainder = " << remainder << "\tnext index = " << remainder - JSCcodingsize * (remainder/JSCcodingsize) << endl; + result_strstrm << JSCcoding[remainder - JSCcodingsize * (remainder/JSCcodingsize)]; + remainder /= JSCcodingsize; + } while (remainder > 0); + + return(result_strstrm.str()); + } + + int Convert_CHAR_to_POSINT (char char_to_convert) + { + // Converts a char into an int according to the coding defined by global constant array JSCcoding. + for (int i = 0; i < JSCcodingsize; ++i) + if (char_to_convert == JSCcoding[i]) return(i); + cout << "char to convert: " << char_to_convert << endl; + JSCerror("Failed to convert char to posint: char not in JSCcoding set."); + return(-1); + } + + int Convert_STR_to_POSINT (string str_to_convert) + { + // Converts a string into a positive integer according to the coding defined by global constant array JSCcoding. + int result = 0; + for (unsigned int i = 0; i < str_to_convert.size(); ++i) { + result = JSCcodingsize * result + Convert_CHAR_to_POSINT(str_to_convert[str_to_convert.size() - 1 - i]); + } + return(result); + } + + + // For reading only the base part of a label: + State_Label_Data Read_Base_Label (string label) + { + // Converts a given label into the appropriate State_Label_Data + + // Split label into base, nexc and q#exc parts: + // these are divided by the two LABELSEP characters in the label. + + string::size_type i1 = label.find(LABELSEP); + string::size_type i2 = label.rfind(LABELSEP); + + string baselabel = label.substr(0, i1); + string nexclabel = label.substr(i1+1, i2-i1-1); + string Ix2exclabel = label.substr(i2+1); + + //cout << "label: " << label << endl; + //cout << "baselabel: " << baselabel << endl; + //cout << "nexclabel: " << nexclabel << endl; + //cout << "Ix2exclabel: " << Ix2exclabel << endl; + + // Read off the base label: count the number of TYPESEP in baselabel + int nbar = 0; + for (unsigned int i = 0; i < baselabel.length(); ++i) if (baselabel[i] == TYPESEP) nbar++; + + //cout << "nbar = " << nbar << endl; + + // There are now nbar + 1 base label data: + int ntypes = nbar + 1; + + //cout << "ntypes = " << ntypes << endl; + + Vect type(ntypes); // integer type labels of the types present + type[0] = 0; // always the case by convention + Vect M(ntypes); // how many particles of each type + + if (ntypes == 1) { // Only one type to read off + istringstream M0buffer(baselabel); + M0buffer >> M[0]; + } + + else { // ntypes > 1 + // Read off M[0]: + string::size_type i1 = baselabel.find(TYPESEP); // M0 is always present, without type specifier + string M0 = baselabel.substr(0, i1); + istringstream M0buffer(M0); + M0buffer >> M[0]; + + // Read off M[1 ... ntypes - 2] + string baselabelremaining = baselabel; + for (int itype = 1; itype < ntypes - 1; ++itype) { + // Remove everything up to leftmost TYPESEP in baselabelremaining + string::size_type i1 = baselabelremaining.find(TYPESEP); // M0 is always present, without type specifier + baselabelremaining = baselabelremaining.substr(i1+1); + string::size_type i2 = baselabelremaining.find(EXCSEP); + string::size_type i3 = baselabelremaining.find(TYPESEP); + string typeread = baselabelremaining.substr(0, i2); + string Mread = baselabelremaining.substr(i2+1,i3-i2-1); + istringstream typereadbuffer (typeread); + typereadbuffer >> type[itype]; + istringstream Mreadbuffer (Mread); + Mreadbuffer >> M[itype]; + } + + // Read off M[ntypes - 1] + { + string::size_type i1 = baselabelremaining.find(TYPESEP); // M0 is always present, without type specifier + baselabelremaining = baselabelremaining.substr(i1+1); + string::size_type i2 = baselabelremaining.find(EXCSEP); + string typeread = baselabelremaining.substr(0, i2); + string Mread = baselabelremaining.substr(i2+1); + istringstream typereadbuffer (typeread); + typereadbuffer >> type[ntypes - 1]; + istringstream Mreadbuffer (Mread); + Mreadbuffer >> M[ntypes - 1]; + } + + } // else if ntypes > 1 + + //cout << "ntypes = " << ntypes << "\tM = " << M << endl; + + // baselabel is now completely read + + // Define some dud nex, Ix2old, Ix2exc: + Vect nexc(ntypes); // how many excitations as compared to the OriginState + Vect > Ix2old(ntypes); // which Ix2 will be excited + Vect > Ix2exc(ntypes); // which Ix2 the excitation has shifted to + + State_Label_Data labeldata (type, M, nexc, Ix2old, Ix2exc); + + return(labeldata); + } + + + + State_Label_Data Read_State_Label (string label, const Vect >& OriginIx2) + { + // Converts a given label into the appropriate State_Label_Data + + // Split label into base, nexc and q#exc parts: + // these are divided by the two LABELSEP characters in the label. + + string::size_type i1 = label.find(LABELSEP); + string::size_type i2 = label.rfind(LABELSEP); + + string baselabel = label.substr(0, i1); + string nexclabel = label.substr(i1+1, i2-i1-1); + string Ix2exclabel = label.substr(i2+1); + + //cout << "label: " << label << endl; + //cout << "baselabel: " << baselabel << endl; + //cout << "nexclabel: " << nexclabel << endl; + //cout << "Ix2exclabel: " << Ix2exclabel << endl; + + // Read off the base label: count the number of TYPESEP in baselabel + int nbar = 0; + for (unsigned int i = 0; i < baselabel.length(); ++i) if (baselabel[i] == TYPESEP) nbar++; + + //cout << "nbar = " << nbar << endl; + + // There are now nbar + 1 base label data: + int ntypes = nbar + 1; + + //cout << "ntypes = " << ntypes << endl; + + Vect type(ntypes); // integer type labels of the types present + type[0] = 0; // always the case by convention + Vect M(ntypes); // how many particles of each type + + if (ntypes == 1) { // Only one type to read off + istringstream M0buffer(baselabel); + M0buffer >> M[0]; + } + + else { // ntypes > 1 + // Read off M[0]: + string::size_type i1 = baselabel.find(TYPESEP); // M0 is always present, without type specifier + string M0 = baselabel.substr(0, i1); + istringstream M0buffer(M0); + M0buffer >> M[0]; + + // Read off M[1 ... ntypes - 2] + string baselabelremaining = baselabel; + for (int itype = 1; itype < ntypes - 1; ++itype) { + // Remove everything up to leftmost TYPESEP in baselabelremaining + string::size_type i1 = baselabelremaining.find(TYPESEP); // M0 is always present, without type specifier + baselabelremaining = baselabelremaining.substr(i1+1); + string::size_type i2 = baselabelremaining.find(EXCSEP); + string::size_type i3 = baselabelremaining.find(TYPESEP); + string typeread = baselabelremaining.substr(0, i2); + string Mread = baselabelremaining.substr(i2+1,i3-i2-1); + istringstream typereadbuffer (typeread); + typereadbuffer >> type[itype]; + istringstream Mreadbuffer (Mread); + Mreadbuffer >> M[itype]; + } + + // Read off M[ntypes - 1] + { + string::size_type i1 = baselabelremaining.find(TYPESEP); // M0 is always present, without type specifier + baselabelremaining = baselabelremaining.substr(i1+1); + string::size_type i2 = baselabelremaining.find(EXCSEP); + string typeread = baselabelremaining.substr(0, i2); + string Mread = baselabelremaining.substr(i2+1); + istringstream typereadbuffer (typeread); + typereadbuffer >> type[ntypes - 1]; + istringstream Mreadbuffer (Mread); + Mreadbuffer >> M[ntypes - 1]; + } + + } // else if ntypes > 1 + + // baselabel is now completely read + + // Read off the nexc vector: + Vect nexc(ntypes); // how many excitations as compared to the OriginState + + if (ntypes == 1) { // Only one type to read off + istringstream exc0buffer(nexclabel); + exc0buffer >> nexc[0]; + } + + else { // ntypes > 1 + // Read off nexc[0]: + string::size_type i1 = nexclabel.find(TYPESEP); + string nexc0 = nexclabel.substr(0, i1); + istringstream nexc0buffer(nexc0); + nexc0buffer >> nexc[0]; + + // Read off nexc[1 ... ntypes - 2] + string nexclabelremaining = nexclabel; + for (int itype = 1; itype < ntypes - 1; ++itype) { + // Remove everything up to leftmost TYPESEP in nexclabelremaining + string::size_type i1 = nexclabelremaining.find(TYPESEP); + nexclabelremaining = nexclabelremaining.substr(i1+1); + string::size_type i2 = nexclabelremaining.find(TYPESEP); + string nexcread = nexclabelremaining.substr(0, i2); + istringstream nexcreadbuffer (nexcread); + nexcreadbuffer >> nexc[itype]; + } + + // Read off nexc[ntypes - 1] + { + string::size_type i1 = nexclabelremaining.find(TYPESEP); + nexclabelremaining = nexclabelremaining.substr(i1+1); + istringstream nexcreadbuffer (nexclabelremaining); + nexcreadbuffer >> nexc[ntypes - 1]; + } + + } // else if ntypes > 1 + + // nexc is now completely read + + //cout << "typefound: " << type << endl; + //cout << "Mfound: " << M << endl; + //cout << "nexcfound: " << nexc << endl; + + + // Now read off the (compressed) jexc and Ix2exc vectors of vectors: + Vect > Ix2old(ntypes); // which Ix2 will be excited + Vect > Ix2exc(ntypes); // which Ix2 the excitation has shifted to + + for (int itype = 0; itype < ntypes; ++itype) { + Ix2old[itype] = Vect (JSC::max(nexc[itype],1)); + Ix2exc[itype] = Vect (JSC::max(nexc[itype],1)); + } + + string Ix2exclabelremaining = Ix2exclabel; + for (int itype = 0; itype < ntypes - 1; ++itype) { + // Read off the Ix2old, Ix2exc: + if (nexc[itype] == 0) { // careful here, need to remove a TYPESEP + string::size_type i2 = Ix2exclabelremaining.find(TYPESEP); + Ix2exclabelremaining = Ix2exclabelremaining.substr(i2+1); + } + for (int iexc = 0; iexc < nexc[itype]; ++iexc) { + //string::size_type i1 = Ix2exclabelremaining.find(INEXCSEP); + string::size_type i2 = (iexc < nexc[itype] - 1 ? Ix2exclabelremaining.find(EXCSEP) : Ix2exclabelremaining.find(TYPESEP)); // careful here! + //cout << "\t" << itype << "\t" << iexc << "\t" << Ix2exclabelremaining << "\t" << i1 << "\t" << i2 << endl; + //string Ix2oldread = Ix2exclabelremaining.substr(0,i1); + //string Ix2excread = Ix2exclabelremaining.substr(i1+1,i2-i1-1); + string Ix2excIDread = Ix2exclabelremaining.substr(0,i2); + int Ix2excID = Convert_STR_to_POSINT(Ix2excIDread); + Ix2old[itype][iexc] = OriginIx2[type[itype] ][Ix2excID - M[itype] * (Ix2excID/M[itype])]; // index is remainder w/r to nr of strings of this type + // Convention: if remainder is even, moving left. If odd, moving right. + // 0 means move one unit left, 1 means move one unit right, etc. + //Ix2exc[itype][iexc] = Ix2old[itype][iexc] + (Ix2excID/M[itype] % 2 ? 2 : -2) * (Ix2excID/(2 * M[itype]) + 0); + Ix2exc[itype][iexc] = Ix2old[itype][iexc] + (Ix2excID/M[itype] % 2 ? 2 : -2) * (Ix2excID/(2 * M[itype]) + 1); // ABACUS++T_8 onwards + + //istringstream Ix2oldreadbuffer (Ix2oldread); + //Ix2oldreadbuffer >> Ix2old[itype][iexc]; + //istringstream Ix2excreadbuffer (Ix2excread); + //Ix2excreadbuffer >> Ix2exc[itype][iexc]; + // Remove everything up to index i2 in Ix2exclabelremaining + Ix2exclabelremaining = Ix2exclabelremaining.substr(i2+1); + } + } + + // Now read off the Ix2old, Ix2exc of the last type: this is always done + for (int iexc = 0; iexc < nexc[ntypes - 1] - 1; ++iexc) { + //string::size_type i1 = Ix2exclabelremaining.find(INEXCSEP); + string::size_type i2 = Ix2exclabelremaining.find(EXCSEP); + //string Ix2oldread = Ix2exclabelremaining.substr(0,i1); + //string Ix2excread = Ix2exclabelremaining.substr(i1+1,i2-i1-1); + string Ix2excIDread = Ix2exclabelremaining.substr(0,i2); + int Ix2excID = Convert_STR_to_POSINT(Ix2excIDread); + Ix2old[ntypes - 1][iexc] = OriginIx2[type[ntypes - 1] ][Ix2excID - M[ntypes - 1] * (Ix2excID/M[ntypes - 1])]; // index is remainder w/r to nr of strings of this type + // Convention: if remainder is even, moving left. If odd, moving right. + //Ix2exc[ntypes - 1][iexc] = Ix2old[ntypes - 1][iexc] + (Ix2excID/M[ntypes - 1] % 2 ? 2 : -2) * (Ix2excID/(2 * M[ntypes - 1]) + 0); + Ix2exc[ntypes - 1][iexc] = Ix2old[ntypes - 1][iexc] + (Ix2excID/M[ntypes - 1] % 2 ? 2 : -2) * (Ix2excID/(2 * M[ntypes - 1]) + 1); // ABACUS++T_8 onwards + + //istringstream Ix2oldreadbuffer (Ix2oldread); + //Ix2oldreadbuffer >> Ix2old[ntypes - 1][iexc]; + //istringstream Ix2excreadbuffer (Ix2excread); + //Ix2excreadbuffer >> Ix2exc[ntypes - 1][iexc]; + // Remove everything up to index i2 in Ix2exclabelremaining + Ix2exclabelremaining = Ix2exclabelremaining.substr(i2+1); + } + + // Now read off the last pair: + int Ix2excID = Convert_STR_to_POSINT(Ix2exclabelremaining); + Ix2old[ntypes - 1][JSC::max(nexc[ntypes - 1] - 1,0)] = OriginIx2[type[ntypes - 1] ][Ix2excID - M[ntypes - 1] * (Ix2excID/M[ntypes - 1])]; // index is remainder w/r to nr of strings of this type + // Convention: if remainder is even, moving left. If odd, moving right. + //Ix2exc[ntypes - 1][JSC::max(nexc[ntypes - 1] - 1,0)] = Ix2old[ntypes - 1][nexc[ntypes - 1] - 1] + (Ix2excID/M[ntypes - 1] % 2 ? 2 : -2) * (Ix2excID/(2 * M[ntypes - 1]) + 0); + Ix2exc[ntypes - 1][JSC::max(nexc[ntypes - 1] - 1,0)] = Ix2old[ntypes - 1][nexc[ntypes - 1] - 1] + (Ix2excID/M[ntypes - 1] % 2 ? 2 : -2) * (Ix2excID/(2 * M[ntypes - 1]) + 1); // ABACUS++T_8 onwards + + //string::size_type ilast = Ix2exclabelremaining.find(INEXCSEP); + //string Ix2oldread = Ix2exclabelremaining.substr(0,ilast); + //string Ix2excread = Ix2exclabelremaining.substr(ilast+1); + //istringstream Ix2oldreadbuffer (Ix2oldread); + //Ix2oldreadbuffer >> Ix2old[ntypes - 1][nexc[ntypes - 1] - 1]; + //istringstream Ix2excreadbuffer (Ix2excread); + //Ix2excreadbuffer >> Ix2exc[ntypes - 1][nexc[ntypes - 1] - 1]; + + //cout << "nexc = " << endl; + //cout << nexc << endl; + //cout << "Ix2old and Ix2exc: " << endl; + //for (int itype = 0; itype < ntypes; ++itype) { + //cout << "Ix2old["<< itype << "]: " << Ix2old[itype] << endl; + //cout << "Ix2exc["<< itype << "]: " << Ix2exc[itype] << endl; + //} + + State_Label_Data labeldata (type, M, nexc, Ix2old, Ix2exc); + + return(labeldata); + } + + State_Label_Data Read_State_Label (string label, const Vect& OriginIx2) + { + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginIx2; + return(Read_State_Label (label, OriginIx2here)); + } + + string Return_State_Label (State_Label_Data data, const Vect >& OriginIx2) + { + // This function produces a compressed label. + + //cout << "Label A" << endl; + + //cout << "\t" << data.M.size() << endl; + //cout << "\t" << data.M[0] << endl; + + string label; + + // Write the base: + // First, particles of type 0: + //cout << "Label A2" << endl; + stringstream M0out; + //cout << "Label A3" << endl; + //cout << "\t" << data.M[0] << endl; + M0out << data.M[0]; + //cout << "Label A4" << endl; + label += M0out.str(); + //cout << "Label A5" << endl; + + for (int itype = 1; itype < data.M.size(); ++itype) { + //cout << "\ta" << itype << "\t" << data.M.size() << endl; + if (data.M[itype] > 0) { + label += TYPESEP; + stringstream typeout; + typeout << data.type[itype]; + label += typeout.str(); + label += EXCSEP; + //cout << "\tc" << endl; + stringstream Mout; + Mout << data.M[itype]; + label += Mout.str(); + //cout << "\td" << endl; + } + } + label += LABELSEP; + + //cout << "Label B" << endl; + + // Now the nexc: + stringstream nexc0out; + nexc0out << data.nexc[0]; + label += nexc0out.str(); + for (int iexc = 1; iexc < data.nexc.size(); ++iexc) { + label += TYPESEP; + stringstream nexcout; + nexcout << data.nexc[iexc]; + label += nexcout.str(); + } + label += LABELSEP; + + //cout << "Label C" << endl; + + // Now the displacements: + // The conventions are as follows. + // For each excitation, an integer number ID is given according to the following rules: + // ID % data.M[itype] gives the index of the hole position in OriginIx2. + // We now define remainder1 == ID - (ID % data.M[itype]). + // remainder1 is interpreted as: remainder1 even/odd means displacement to left/right + // remainder1/2 + 1 gives then the displacement in units of quantum nr. + // The +1 is to start labeling displacement from 0 (so 0 means displace by one unit). + for (int itype = 0; itype < data.M.size(); ++itype) { + if (itype > 0) label += TYPESEP; + for (int iexc = 0; iexc < data.nexc[itype]; ++iexc) { + if (iexc > 0) label += EXCSEP; + int excID = abs(data.Ix2exc[itype][iexc] - data.Ix2old[itype][iexc]) - 2; // necessarily even and >= 0 + if (data.Ix2exc[itype][iexc] > data.Ix2old[itype][iexc]) excID += 1; // make odd if displacement is to the right + int holeindex = -1; + do { + holeindex++; + } while (OriginIx2[data.type[itype] ][holeindex] != data.Ix2old[itype][iexc] && holeindex < OriginIx2[data.type[itype] ].size() - 1); + if (holeindex == OriginIx2[data.type[itype] ].size()) JSCerror("Going out of bounds in Compress_Label."); + excID = excID * data.M[itype] + holeindex; + label += Convert_POSINT_to_STR(excID); + } // for iexc + } // for itype + + //cout << "Label D" << endl; + + return(label); + } + + string Return_State_Label (State_Label_Data data, const Vect& OriginIx2) + { + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginIx2; + return(Return_State_Label (data, OriginIx2here)); + } + + string Return_State_Label (const Vect >& ScanIx2, const Vect >& OriginIx2) + { + // This function does not assume any ordering of the Ix2. + + if (ScanIx2.size() != OriginIx2.size()) JSCerror("ScanIx2.size() != OriginIx2.size() in Find_Label."); + for (int i = 0; i < ScanIx2.size(); ++i) if (ScanIx2[i].size() != OriginIx2[i].size()) JSCerror("ScanIx2[i].size() != OriginIx2[i].size() in Find_Label."); + + // Set the state ulabel: + // Count the number of types present: + int ntypespresent = 0; + for (int is = 0; is < ScanIx2.size(); ++is) if (is == 0 || ScanIx2[is].size() > 0) ntypespresent++; // type 0 is by default always present + + Vect type_ref(ntypespresent); + Vect M_ref(ntypespresent); + Vect nexc_ref(0, ntypespresent); + + // Define type_ref and M_ref: + int ntypespresentcheck = 0; + for (int is = 0; is < ScanIx2.size(); ++is) if (is == 0 || ScanIx2[is].size() > 0) { // type 0 is by default always present + type_ref[ntypespresentcheck] = is; + M_ref[ntypespresentcheck++] = ScanIx2[is].size(); + } + if (ntypespresentcheck != ntypespresent) JSCerror("Counting types present wrong in Return_Label."); + + // Count nr of particle-holes: + for (int it = 0; it < ntypespresent; ++it) + for (int i = 0; i < M_ref[it]; ++i) if (!OriginIx2[type_ref[it] ].includes(ScanIx2[type_ref[it] ][i])) nexc_ref[it] += 1; + Vect > Ix2old_ref(ntypespresent); + Vect > Ix2exc_ref(ntypespresent); + for (int it = 0; it < ntypespresent; ++it) Ix2old_ref[it] = Vect(JSC::max(nexc_ref[it],1)); + for (int it = 0; it < ntypespresent; ++it) Ix2exc_ref[it] = Vect(JSC::max(nexc_ref[it],1)); + for (int it = 0; it < ntypespresent; ++it) { + int nexccheck = 0; + for (int i = 0; i < M_ref[it]; ++i) if (!OriginIx2[type_ref[it] ].includes(ScanIx2[type_ref[it] ][i])) Ix2exc_ref[it][nexccheck++] = ScanIx2[type_ref[it] ][i]; + if (nexccheck != nexc_ref[it]) JSCerror("Counting excitations wrong (1) in Return_State_Label"); + nexccheck = 0; + for (int i = 0; i < M_ref[it]; ++i) if (!ScanIx2[type_ref[it] ].includes(OriginIx2[type_ref[it] ][i])) Ix2old_ref[it][nexccheck++] = OriginIx2[type_ref[it] ][i]; + if (nexccheck != nexc_ref[it]) { + cout << OriginIx2 << endl; + cout << ScanIx2 << endl; + cout << nexc_ref[it] << endl; + cout << Ix2exc_ref[it] << endl; + JSCerror("Counting excitations wrong (2) in Return_State_Label"); + } + // Now order the Ix2old_ref and Ix2exc_ref: + Ix2old_ref[it].QuickSort(); + Ix2exc_ref[it].QuickSort(); + } // for it + + State_Label_Data labeldata(type_ref, M_ref, nexc_ref, Ix2old_ref, Ix2exc_ref); + + return(Return_State_Label (labeldata, OriginIx2)); + } + + string Return_State_Label (const Vect& ScanIx2, const Vect& OriginIx2) + { + Vect > ScanIx2here(1); + ScanIx2here[0] = ScanIx2; + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginIx2; + return(Return_State_Label (ScanIx2here, OriginIx2here)); + } + + + //bool Set_to_Label (string label_ref, Vect >& Ix2, const Vect >& OriginIx2) + Vect > Return_Ix2_from_Label (string label_ref, const Vect >& OriginIx2) + { + // ASSUMPTIONS: + // OriginIx2 is ordered. + + //Ix2 = OriginIx2; // this will fail if the sizes are incompatible + Vect > Ix2 = OriginIx2; // this will fail if the sizes are incompatible + + State_Label_Data labeldata = Read_State_Label (label_ref, OriginIx2); + + //cout << "Read label OK" << endl; + + // Now set the excitations: + for (int it = 0; it < labeldata.type.size(); ++it) + for (int iexc = 0; iexc < labeldata.nexc[it]; ++iexc) + for (int i = 0; i < labeldata.M[it]; ++i) + if (Ix2[labeldata.type[it] ][i] == labeldata.Ix2old[it][iexc]) { + Ix2[labeldata.type[it] ][i] = labeldata.Ix2exc[it][iexc]; + } + + // Now reorder the Ix2 to follow convention: + for (int il = 0; il < Ix2.size(); ++il) Ix2[il].QuickSort(); + + //cout << "Ix2 found = " << Ix2 << endl; + + return(Ix2); + //return(true); + } + // Specialization to Lieb-Liniger: + Vect Return_Ix2_from_Label (string label_ref, const Vect& OriginIx2) + { + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginIx2; + + return(Return_Ix2_from_Label(label_ref, OriginIx2here)[0]); + } + + + /* DEPRECATED ++G_5 + // Conversion from one to the other: + string Compress_Ulabel (string ulabel_ref, const Vect >& OriginIx2) + { + // From a normal label, return a compressed one. + State_Label_Data data = Read_State_Ulabel (ulabel_ref); + + return(Return_State_Label(data, OriginIx2)); + } + + string Compress_Ulabel (string ulabel_ref, const Vect& OriginIx2) + // if there is only one type + { + // From a normal label, return a compressed one. + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginIx2; + + State_Label_Data data = Read_State_Ulabel (ulabel_ref); + + return(Return_State_Label(data, OriginIx2here)); + } + + string Uncompress_Label (string label_ref, const Vect >& OriginIx2) + { + // From a compressed label, return a normal one. + State_Label_Data data = Read_State_Label (label_ref, OriginIx2); + + return(Return_State_Ulabel(data)); + } + + string Uncompress_Label (string label_ref, const Vect& OriginIx2) + // if there is only one type + { + // From a compressed label, return a normal one. + Vect > OriginIx2here(1); + OriginIx2here[0] = OriginIx2; + + State_Label_Data data = Read_State_Label (label_ref, OriginIx2here); + + return(Return_State_Ulabel(data)); + } + */ + +} // namespace JSC diff --git a/src/XXX_h0/SF_4p_client.cc b/src/XXX_h0/SF_4p_client.cc new file mode 100755 index 0000000..a204f7f --- /dev/null +++ b/src/XXX_h0/SF_4p_client.cc @@ -0,0 +1,172 @@ +/********************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) 2006. + +----------------------------------------------------------- + +File: SF_4p_client.cc + +Purpose: 4 spinon contrib to S(k, omega) for XXX_h0, parallel code using MPICH. + Client part. + +Last modified: 13/10/06 + +***********************************************************/ + +#include "mpi.h" +#include "JSC.h" + +namespace JSC { + + void SF_4p_opt_client (MPI_Comm comm, I_table Itable) + { + int server = 0; + int rank; + + MPI_Status status; + + int client_request_size = 7; + DP client_request[client_request_size]; + // this is: + // ik + // iomega + // k + // omega + // req_prec + // Npts_K + // Npts_W + int client_result_size = 8; + DP client_result[client_result_size]; // as above, + SF_4p + + MPI_Comm_rank (comm, &rank); + + MPI_Recv (client_request, client_request_size, MPI_DOUBLE, server, MPI_ANY_TAG, comm, &status); + //cout << "Client of rank " << rank << ": received request "; + //for (int i = 0; i < client_request_size; ++i) cout << client_request[i] << " "; + //cout << "with tag " << status.MPI_TAG << endl; + + + while (status.MPI_TAG) { + + // Result: SF_4p_opt (k, omega) + client_result[0] = client_request[0]; + client_result[1] = client_request[1]; + client_result[2] = client_request[2]; + client_result[3] = client_request[3]; + client_result[4] = client_request[4]; + client_result[5] = client_request[5]; + client_result[6] = client_request[6]; + client_result[7] = SF_4p_opt (client_request[2], client_request[3], client_request[4], + int(client_request[5]), int(client_request[6]), Itable); + MPI_Send (client_result, client_result_size, MPI_DOUBLE, server, status.MPI_TAG, comm); + //cout << "Client of rank " << rank << " sending complete" << endl; + + // Wait for subsequent request from server + MPI_Recv (client_request, client_request_size, MPI_DOUBLE, server, MPI_ANY_TAG, comm, &status); + //cout << "Client of rank " << rank << ": received request "; + //for (int i = 0; i < client_request_size; ++i) cout << client_request[i] << " "; + //cout << "with tag " << status.MPI_TAG << endl; + + } + + return; + } + /* + void SF_4p_kwKW_alpha_client (MPI_Comm comm, DP req_prec, int max_rec, I_table Itable) + { + int server = 0; + int rank; + + MPI_Status status; + + int client_request_size = 2; + DP client_request[client_request_size]; // this is k, alpha + int client_result_size = 3; + DP client_result[client_result_size]; // this is k, alpha, SF_4p_kwKW_alpha + + MPI_Comm_rank (comm, &rank); + + MPI_Recv (client_request, client_request_size, MPI_DOUBLE, server, MPI_ANY_TAG, comm, &status); + //cout << "Client of rank " << rank << ": received request "; + //for (int i = 0; i < client_request_size; ++i) cout << client_request[i] << " "; + //cout << "with tag " << status.MPI_TAG << endl; + + Vect_DP args_to_SF_4p_kwKW_alpha(4); + args_to_SF_4p_kwKW_alpha[2] = req_prec; + args_to_SF_4p_kwKW_alpha[3] = DP(max_rec); + + while (status.MPI_TAG) { + + args_to_SF_4p_kwKW_alpha[0] = client_request[0]; + args_to_SF_4p_kwKW_alpha[1] = client_request[1]; + + // Result: SF_4p (k, omega) + client_result[0] = client_request[0]; + client_result[1] = client_request[1]; + client_result[2] = SF_4p_kwKW_alpha (args_to_SF_4p_kwKW_alpha, Itable); + + MPI_Send (client_result, client_result_size, MPI_DOUBLE, server, status.MPI_TAG, comm); + //cout << "Client of rank " << rank << " sending complete" << endl; + + // Wait for subsequent request from server + MPI_Recv (client_request, client_request_size, MPI_DOUBLE, server, MPI_ANY_TAG, comm, &status); + //cout << "Client of rank " << rank << ": received request "; + //for (int i = 0; i < client_request_size; ++i) cout << client_request[i] << " "; + //cout << "with tag " << status.MPI_TAG << endl; + + } + + return; + } + + void SF_4p_kwKW_alpha_opt_client (MPI_Comm comm, DP req_prec, int Npts_K, int Npts_W, I_table Itable) + { + int server = 0; + int rank; + + MPI_Status status; + + int client_request_size = 2; + DP client_request[client_request_size]; // this is k, alpha + int client_result_size = 3; + DP client_result[client_result_size]; // this is k, alpha, SF_4p_kwKW_alpha_opt + + MPI_Comm_rank (comm, &rank); + + MPI_Recv (client_request, client_request_size, MPI_DOUBLE, server, MPI_ANY_TAG, comm, &status); + //cout << "Client of rank " << rank << ": received request "; + //for (int i = 0; i < client_request_size; ++i) cout << client_request[i] << " "; + //cout << "with tag " << status.MPI_TAG << endl; + + Vect_DP args_to_SF_4p_kwKW_alpha_opt(5); + args_to_SF_4p_kwKW_alpha_opt[2] = req_prec; + args_to_SF_4p_kwKW_alpha_opt[3] = DP(Npts_K); + args_to_SF_4p_kwKW_alpha_opt[4] = DP(Npts_W); + + while (status.MPI_TAG) { + + args_to_SF_4p_kwKW_alpha_opt[0] = client_request[0]; + args_to_SF_4p_kwKW_alpha_opt[1] = client_request[1]; + + // Result: SF_4p (k, omega) + client_result[0] = client_request[0]; + client_result[1] = client_request[1]; + client_result[2] = SF_4p_kwKW_alpha_opt (args_to_SF_4p_kwKW_alpha_opt, Itable); + + MPI_Send (client_result, client_result_size, MPI_DOUBLE, server, status.MPI_TAG, comm); + //cout << "Client of rank " << rank << " sending complete" << endl; + + // Wait for subsequent request from server + MPI_Recv (client_request, client_request_size, MPI_DOUBLE, server, MPI_ANY_TAG, comm, &status); + //cout << "Client of rank " << rank << ": received request "; + //for (int i = 0; i < client_request_size; ++i) cout << client_request[i] << " "; + //cout << "with tag " << status.MPI_TAG << endl; + + } + + return; + } + */ +} // namespace JSC diff --git a/src/XXX_h0/SF_4p_server.cc b/src/XXX_h0/SF_4p_server.cc new file mode 100755 index 0000000..f592d1c --- /dev/null +++ b/src/XXX_h0/SF_4p_server.cc @@ -0,0 +1,572 @@ +/********************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) 2006. + +----------------------------------------------------------- + +File: SF_4p_server.cc + +Purpose: 4 spinon contrib to S(k, omega) for XXX_h0, parallel code using MPICH. + Server part. + +Last modified: 13/10/06 + +***********************************************************/ + +#include "mpi.h" +#include "JSC.h" + +namespace JSC { + + + void SF_4p_opt_server (MPI_Comm comm, int N, DP omegamax, int Nomega, DP req_prec, int Npts_K, int Npts_W, I_table Itable, + int Max_Secs, bool refine) + { + double Start_Time_MPI = MPI_Wtime(); + + Heis_Write_Kfull_File (N); + Heis_Write_w_File (Nomega, omegamax); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_4p_N_" << N << "_No_" << Nomega << "_omax_" << omegamax + << "_prec_" << req_prec << "_Npts_K_" << Npts_K << "_Npts_W_" << Npts_W << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + fstream SF_outfile; + if (!refine) SF_outfile.open(SF_Cstr, fstream::out | fstream::trunc); + else SF_outfile.open(SF_Cstr, fstream::in | fstream::out); + if (SF_outfile.fail()) JSCerror("Could not open SF_outfile... "); + + SF_outfile.precision(10); + + int dim_k = N/2 + 1; + + DP* SF_4p = new DP[dim_k * Nomega]; + + // Initialize to zero to be sure: + for (int i = 0; i < Nomega * dim_k; ++i) SF_4p[i] = 0.0; + + // If refining, load SF from existing file + + int total_nr_req = 0; + DP buff; + if (refine) { + for (int iomega = 0; iomega < Nomega; ++iomega) { + for (int ik = 0; ik < dim_k; ++ik) { + SF_outfile >> SF_4p[dim_k * iomega + ik]; + if (SF_4p[dim_k * iomega + ik] == 0.0) total_nr_req++; + //cout << ik << "\t" << iomega << "\t" << SF_4p[dim_k * iomega + ik] << "\t" << (SF_4p[dim_k * iomega + ik] == 0.0) << "\t" << total_nr_req << endl; + + // We only load the LHS of the BZ, so we load N/2 empty values... + } + for (int ibuff = 0; ibuff < N/2; ++ibuff) SF_outfile >> buff; + } + } + else if (!refine) total_nr_req = dim_k * Nomega; + + // List iomega and ik which need to be computed: + + int* needed_ik = new int[total_nr_req]; + int* needed_iomega = new int[total_nr_req]; + + int index = 0; + for (int iomega = 0; iomega < Nomega; ++iomega) + for (int ik = 0; ik < dim_k; ++ik) { + if (SF_4p[dim_k * iomega + ik] == 0) { + needed_ik[index] = ik; + needed_iomega[index] = iomega; + index++; + } + } + + cout << total_nr_req << " points required." << endl; + + if (index != total_nr_req) { + cout << total_nr_req << "\t" << index << endl; + JSCerror("Not counting total_nr_req correctly in SF_4p_opt_server"); + } + //ofstream SFsrc_outfile; + //SFsrc_outfile.open(SFsrc_Cstr); + + //DP omegamin_used = 0.5 * wmin_4p (k); // Correct for factor of 2 in E between me & Bougourzi + //DP omegamax_used = 0.5 * wmax_4p (k); + + int nr_machines; + MPI_Comm_size (comm, &nr_machines); + int nr_clients = nr_machines - 1; // one for the server + + int client_request_size = 7; + DP client_request[client_request_size]; + // this is: + // ik + // iomega + // k + // omega + // req_prec + // Npts_K + // Npts_W + int client_result_size = 8; + DP client_result[client_result_size]; // same, plus SF_4p + + MPI_Status status; + + int scanning = 1; + + int nr_sent_out = 0; + int nr_returned = 0; + + for (int i = 1; i <= nr_clients && i <= total_nr_req; ++i) { + + // Send request to client i, in the form of the req_id_array vector + client_request[0] = DP(needed_ik[nr_sent_out]); + client_request[1] = DP(needed_iomega[nr_sent_out]); + client_request[2] = (twoPI * needed_ik[nr_sent_out])/N; + client_request[3] = (needed_iomega[nr_sent_out] + 0.5) * omegamax/Nomega; + client_request[4] = req_prec; + client_request[5] = DP(Npts_K); + client_request[6] = DP(Npts_W); + + MPI_Send(client_request, client_request_size, MPI_DOUBLE, i, scanning, comm); + //cout << "Server: sent request "; + //for (int ii = 0; ii < client_request_size; ++ii) cout << client_request[ii] << " "; + //cout << "with tag " << scanning << " to client " << i << endl; + nr_sent_out++; + //cout << "nr_sent_out = " << nr_sent_out << endl; + } + + DP Actual_Time_MPI = MPI_Wtime(); + + //while (nr_returned < total_nr_req) { + while (nr_returned < nr_sent_out) { + + //cout << "Server: waiting for answers... " << endl; + MPI_Recv (client_result, client_result_size, MPI_DOUBLE, MPI_ANY_SOURCE, MPI_ANY_TAG, comm, &status); + nr_returned++; + //cout << "Server: received answer from client " << status.MPI_SOURCE << endl; + // unbuffer result + SF_4p[dim_k * int(client_result[1]) + int(client_result[0])] = client_result[7]; + + Actual_Time_MPI = MPI_Wtime(); + + // Send out new request if needed and time available + if (nr_sent_out < total_nr_req && Actual_Time_MPI - Start_Time_MPI < Max_Secs) { + client_request[0] = needed_ik[nr_sent_out]; + client_request[1] = needed_iomega[nr_sent_out]; + client_request[2] = (twoPI * needed_ik[nr_sent_out])/N; + client_request[3] = (needed_iomega[nr_sent_out] + 0.5) * omegamax/Nomega; + MPI_Send (client_request, client_request_size, MPI_DOUBLE, status.MPI_SOURCE, scanning, comm); + nr_sent_out++; + //cout << "nr_sent_out = " << nr_sent_out << endl; + } + + } // while (nr_returned < total_nr_req) + + cout << endl << "Computed " << nr_returned << " points ouf of " << total_nr_req << " required. " << endl; + + cout << endl << "Server saving file..." << endl; + + // Output all data: double up to full [0, 2pi] interval in k with symmetry + SF_outfile.seekp(0); + for (int iomega = 0; iomega < Nomega; ++iomega) { + for (int ik = 0; ik < dim_k; ++ik) SF_outfile << SF_4p[dim_k * iomega + ik] << "\t"; + for (int ik = 2; ik <= dim_k; ++ik) SF_outfile << SF_4p[dim_k * iomega + dim_k - ik] << "\t"; + SF_outfile << endl; + } + + SF_outfile.close(); + + cout << endl << "Done !" << endl; + + // Send term signal to clients + int scanning_completed = 0; + for (int i = 1; i <= nr_clients; ++i) + MPI_Send (client_request, client_request_size, MPI_DOUBLE, i, scanning_completed, comm); + + return; + } + + /**************************************************** + + // TO COMPLETE FOR 5.5 (commented out in 5.7) + + *********************************************************/ + /* + void SF_4p_opt_server (MPI_Comm comm, DP k_needed, DP omegamax, int Nomega, DP req_prec, int Npts_K, int Npts_W, I_table Itable, + int Max_Secs, bool refine) + { + double Start_Time_MPI = MPI_Wtime(); + + Heis_Write_w_File (Nomega, omegamax); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_4p_k_" << k_needed << "_No_" << Nomega << "_omax_" << omegamax + << "_prec_" << req_prec << "_Npts_K_" << Npts_K << "_Npts_W_" << Npts_W << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + fstream SF_outfile; + if (!refine) SF_outfile.open(SF_Cstr, fstream::out | fstream::trunc); + else SF_outfile.open(SF_Cstr, fstream::in | fstream::out); + if (SF_outfile.fail()) JSCerror("Could not open SF_outfile... "); + + SF_outfile.precision(12); + + int dim_k = 1; + + DP* SF_4p = new DP[dim_k * Nomega]; + + // Initialize to zero to be sure: + for (int i = 0; i < Nomega * dim_k; ++i) SF_4p[i] = 0.0; + + // If refining, load SF from existing file + + int total_nr_req = 0; + DP buff; + if (refine) { + for (int iomega = 0; iomega < Nomega; ++iomega) { + for (int ik = 0; ik < dim_k; ++ik) { + SF_outfile >> SF_4p[dim_k * iomega + ik]; + if (SF_4p[dim_k * iomega + ik] == 0.0) total_nr_req++; + //cout << ik << "\t" << iomega << "\t" << SF_4p[dim_k * iomega + ik] << "\t" << (SF_4p[dim_k * iomega + ik] == 0.0) << "\t" << total_nr_req << endl; + + // We only load the LHS of the BZ, so we load N/2 empty values... + } + for (int ibuff = 0; ibuff < N/2; ++ibuff) SF_outfile >> buff; + } + } + else if (!refine) total_nr_req = dim_k * Nomega; + + // List iomega and ik which need to be computed: + + int* needed_ik = new int[total_nr_req]; + int* needed_iomega = new int[total_nr_req]; + + int index = 0; + for (int iomega = 0; iomega < Nomega; ++iomega) + for (int ik = 0; ik < dim_k; ++ik) { + if (SF_4p[dim_k * iomega + ik] == 0) { + needed_ik[index] = ik; + needed_iomega[index] = iomega; + index++; + } + } + + cout << total_nr_req << " points required." << endl; + + if (index != total_nr_req) { + cout << total_nr_req << "\t" << index << endl; + JSCerror("Not counting total_nr_req correctly in SF_4p_opt_server"); + } + //ofstream SFsrc_outfile; + //SFsrc_outfile.open(SFsrc_Cstr); + + //DP omegamin_used = 0.5 * wmin_4p (k); // Correct for factor of 2 in E between me & Bougourzi + //DP omegamax_used = 0.5 * wmax_4p (k); + + int nr_machines; + MPI_Comm_size (comm, &nr_machines); + int nr_clients = nr_machines - 1; // one for the server + + int client_request_size = 7; + DP client_request[client_request_size]; + // this is: + // ik + // iomega + // k + // omega + // req_prec + // Npts_K + // Npts_W + int client_result_size = 8; + DP client_result[client_result_size]; // same, plus SF_4p + + MPI_Status status; + + int scanning = 1; + + int nr_sent_out = 0; + int nr_returned = 0; + + for (int i = 1; i <= nr_clients && i <= total_nr_req; ++i) { + + // Send request to client i, in the form of the req_id_array vector + client_request[0] = DP(needed_ik[nr_sent_out]); + client_request[1] = DP(needed_iomega[nr_sent_out]); + client_request[2] = k_needed;//(twoPI * needed_ik[nr_sent_out])/N; + client_request[3] = (needed_iomega[nr_sent_out] + 0.5) * omegamax/Nomega; + client_request[4] = req_prec; + client_request[5] = DP(Npts_K); + client_request[6] = DP(Npts_W); + + MPI_Send(client_request, client_request_size, MPI_DOUBLE, i, scanning, comm); + //cout << "Server: sent request "; + //for (int ii = 0; ii < client_request_size; ++ii) cout << client_request[ii] << " "; + //cout << "with tag " << scanning << " to client " << i << endl; + nr_sent_out++; + //cout << "nr_sent_out = " << nr_sent_out << endl; + } + + DP Actual_Time_MPI = MPI_Wtime(); + + //while (nr_returned < total_nr_req) { + while (nr_returned < nr_sent_out) { + + //cout << "Server: waiting for answers... " << endl; + MPI_Recv (client_result, client_result_size, MPI_DOUBLE, MPI_ANY_SOURCE, MPI_ANY_TAG, comm, &status); + nr_returned++; + //cout << "Server: received answer from client " << status.MPI_SOURCE << endl; + // unbuffer result + SF_4p[int(client_result[1])] = client_result[7]; + + Actual_Time_MPI = MPI_Wtime(); + + // Send out new request if needed and time available + if (nr_sent_out < total_nr_req && Actual_Time_MPI - Start_Time_MPI < Max_Secs) { + client_request[0] = needed_ik[nr_sent_out]; + client_request[1] = needed_iomega[nr_sent_out]; + client_request[2] = k_needed;//(twoPI * needed_ik[nr_sent_out])/N; + client_request[3] = (needed_iomega[nr_sent_out] + 0.5) * omegamax/Nomega; + MPI_Send (client_request, client_request_size, MPI_DOUBLE, status.MPI_SOURCE, scanning, comm); + nr_sent_out++; + //cout << "nr_sent_out = " << nr_sent_out << endl; + } + + } // while (nr_returned < total_nr_req) + + cout << endl << "Computed " << nr_returned << " points ouf of " << total_nr_req << " required. " << endl; + + cout << endl << "Server saving file..." << endl; + + // Output all data: + SF_outfile.seekp(0); + for (int iomega = 0; iomega < Nomega; ++iomega) { + SF_outfile << omega[iomega] << "\t" << SF_4p[iomega] << endl; + } + + SF_outfile.close(); + + cout << endl << "Done !" << endl; + + // Send term signal to clients + int scanning_completed = 0; + for (int i = 1; i <= nr_clients; ++i) + MPI_Send (client_request, client_request_size, MPI_DOUBLE, i, scanning_completed, comm); + + return; + } + */ + /* + // Function producing a fixed k scan, with data file + void SF_4p_rec_par (MPI_Comm comm, DP k, DP req_prec, int max_rec_w, int max_rec, I_table Itable) + { + int Npts_w = int(pow(3.0, max_rec_w + 2)); + + stringstream SFraw_stringstream; + string SFraw_string; + SFraw_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_max_rec_w_" << max_rec_w << "_max_rec_" << max_rec << ".raw"; + SFraw_string = SFraw_stringstream.str(); + const char* SFraw_Cstr = SFraw_string.c_str(); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_max_rec_w_" << max_rec_w << "_max_rec_" << max_rec << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + stringstream SFsrc_stringstream; + string SFsrc_string; + SFsrc_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_max_rec_w_" << max_rec_w << "_max_rec_" << max_rec << ".src"; + SFsrc_string = SFsrc_stringstream.str(); + const char* SFsrc_Cstr = SFsrc_string.c_str(); + + ofstream SFraw_outfile; + SFraw_outfile.open(SFraw_Cstr); + //ofstream SFsrc_outfile; + //SFsrc_outfile.open(SFsrc_Cstr); + + //DP omegamin_used = 0.5 * wmin_4p (k); // Correct for factor of 2 in E between me & Bougourzi + //DP omegamax_used = 0.5 * wmax_4p (k); + + int nr_machines; + MPI_Comm_size (comm, &nr_machines); + int nr_clients = nr_machines - 1; // one for the server + + int client_request_size = 2; + DP client_request[client_request_size]; // this is k, alpha + int client_result_size = 3; + DP client_result[client_result_size]; // this is k, alpha, SF_4p_kwKW_alpha + + MPI_Status status; + + int scanning = 1; + + int total_nr_req = Npts_w; + int nr_sent_out = 0; + int nr_returned = 0; + + Vect_DP alpha_req(Npts_w); + for (int iw = 0; iw < Npts_w; ++iw) alpha_req[iw] = 0.5 * PI * (iw + 0.5)/Npts_w; + + //cout << "alpha_req = " << alpha_req << endl; + + for (int i = 1; i <= nr_clients && i <= total_nr_req; ++i) { + + // Send request to client i, in the form of the req_id_array vector + client_request[0] = k; + client_request[1] = alpha_req[nr_sent_out]; + + MPI_Send(client_request, client_request_size, MPI_DOUBLE, i, scanning, comm); + //cout << "Server: sent request "; + //for (int ii = 0; ii < client_request_size; ++ii) cout << client_request[ii] << " "; + //cout << "with tag " << scanning << " to client " << i << endl; + nr_sent_out++; + //cout << "nr_sent_out = " << nr_sent_out << endl; + } + + while (nr_returned < total_nr_req) { + + //cout << "Server: waiting for answers... " << endl; + MPI_Recv (client_result, client_result_size, MPI_DOUBLE, MPI_ANY_SOURCE, MPI_ANY_TAG, comm, &status); + nr_returned++; + //cout << "Server: received answer from client " << status.MPI_SOURCE << endl; + + // unbuffer result + SFraw_outfile << client_result[1] << "\t" << client_result[2] << endl; + + // Send out new request if needed + if (nr_sent_out < total_nr_req) { + client_request[1] = alpha_req[nr_sent_out]; + MPI_Send (client_request, client_request_size, MPI_DOUBLE, status.MPI_SOURCE, scanning, comm); + nr_sent_out++; + //cout << "nr_sent_out = " << nr_sent_out << endl; + } + + } // while (nr_returned < total_nr_req) + + // Send term signal to clients + int scanning_completed = 0; + for (int i = 1; i <= nr_clients; ++i) + MPI_Send (client_request, client_request_size, MPI_DOUBLE, i, scanning_completed, comm); + + SFraw_outfile.close(); + + //SFsrc_outfile << answer << endl; + //SFsrc_outfile.close(); + + // Translate raw data into SF_4p (k,omega) data + + //Translate_raw_4p_data (k, max_rec_w, SFraw_Cstr, SF_Cstr, SFsrc_Cstr, Itable); + Translate_raw_4p_data (k, int(pow(3.0, max_rec_w + 2)), SFraw_Cstr, SF_Cstr, SFsrc_Cstr, Itable); + + return; + } + + void SF_4p_opt_par (MPI_Comm comm, DP k, DP req_prec, int Npts_w, int Npts_K, int Npts_W, I_table Itable) + { + stringstream SFraw_stringstream; + string SFraw_string; + SFraw_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << ".raw"; + SFraw_string = SFraw_stringstream.str(); + const char* SFraw_Cstr = SFraw_string.c_str(); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + stringstream SFsrc_stringstream; + string SFsrc_string; + SFsrc_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << ".src"; + SFsrc_string = SFsrc_stringstream.str(); + const char* SFsrc_Cstr = SFsrc_string.c_str(); + + ofstream SFraw_outfile; + SFraw_outfile.open(SFraw_Cstr); + //ofstream SFsrc_outfile; + //SFsrc_outfile.open(SFsrc_Cstr); + + //DP omegamin_used = 0.5 * wmin_4p (k); // Correct for factor of 2 in E between me & Bougourzi + //DP omegamax_used = 0.5 * wmax_4p (k); + + int nr_machines; + MPI_Comm_size (comm, &nr_machines); + int nr_clients = nr_machines - 1; // one for the server + + int client_request_size = 2; + DP client_request[client_request_size]; // this is k, alpha + int client_result_size = 3; + DP client_result[client_result_size]; // this is k, alpha, SF_4p_kwKW_alpha_opt + + MPI_Status status; + + int scanning = 1; + + int total_nr_req = Npts_w; + int nr_sent_out = 0; + int nr_returned = 0; + + Vect_DP alpha_req(Npts_w); + for (int iw = 0; iw < Npts_w; ++iw) alpha_req[iw] = 0.5 * PI * (iw + 0.5)/Npts_w; + + //cout << "alpha_req = " << alpha_req << endl; + + for (int i = 1; i <= nr_clients && i <= total_nr_req; ++i) { + + // Send request to client i, in the form of the req_id_array vector + client_request[0] = k; + client_request[1] = alpha_req[nr_sent_out]; + + MPI_Send(client_request, client_request_size, MPI_DOUBLE, i, scanning, comm); + //cout << "Server: sent request "; + //for (int ii = 0; ii < client_request_size; ++ii) cout << client_request[ii] << " "; + //cout << "with tag " << scanning << " to client " << i << endl; + nr_sent_out++; + //cout << "nr_sent_out = " << nr_sent_out << endl; + } + + while (nr_returned < total_nr_req) { + + //cout << "Server: waiting for answers... " << endl; + MPI_Recv (client_result, client_result_size, MPI_DOUBLE, MPI_ANY_SOURCE, MPI_ANY_TAG, comm, &status); + nr_returned++; + //cout << "Server: received answer from client " << status.MPI_SOURCE << endl; + + // unbuffer result + SFraw_outfile << client_result[1] << "\t" << client_result[2] << endl; + + // Send out new request if needed + if (nr_sent_out < total_nr_req) { + client_request[1] = alpha_req[nr_sent_out]; + MPI_Send (client_request, client_request_size, MPI_DOUBLE, status.MPI_SOURCE, scanning, comm); + nr_sent_out++; + //cout << "nr_sent_out = " << nr_sent_out << endl; + } + + } // while (nr_returned < total_nr_req) + + // Send term signal to clients + int scanning_completed = 0; + for (int i = 1; i <= nr_clients; ++i) + MPI_Send (client_request, client_request_size, MPI_DOUBLE, i, scanning_completed, comm); + + SFraw_outfile.close(); + + //SFsrc_outfile << answer << endl; + //SFsrc_outfile.close(); + + // Translate raw data into SF_4p (k,omega) data + + //Translate_raw_4p_data (k, max_rec_w, SFraw_Cstr, SF_Cstr, SFsrc_Cstr, Itable); + Translate_raw_4p_data (k, Npts_w, SFraw_Cstr, SF_Cstr, SFsrc_Cstr, Itable); + + return; + } + */ + +} // namespace JSC diff --git a/src/XXX_h0/XXX_h0.cc b/src/XXX_h0/XXX_h0.cc new file mode 100755 index 0000000..8e3ad75 --- /dev/null +++ b/src/XXX_h0/XXX_h0.cc @@ -0,0 +1,1752 @@ +/**************************************************************** + +XXX_h0.cc + +Defines all class procedures used for the XXX chain in zero field. + +LAST MODIFIED: 12/10/06 + +******************************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + DP I_integral (DP rho, DP req_prec) + { + DP t1 = 1.0; // delimiter between two integrals + DP tinf = 24.0; // such that exp(-2tinf) << machine_eps + DP Euler_Mascheroni = 0.577215664901532860606; + + DP rho_used = fabs(rho); + + if (rho_used > 10000.0) return(-PI * rho_used - 2.0 * Euler_Mascheroni); // CHECK THIS + + Vect_DP args(2); + args[0] = 0.0; + args[1] = rho_used; + + DP answer = -2.0 * Euler_Mascheroni - 2.0 * log(4.0 * rho_used * t1) + + 2.0 * Integrate_rec (Integrand_11, args, 0, 0.0, t1, req_prec, 12) + - 2.0 * Integrate_rec (Integrand_12, args, 0, t1, tinf, req_prec, 12) + - Integrate_rec (Integrand_2, args, 0, 0.0, tinf, req_prec, 12); + + /* + DP answer = -2.0 * Euler_Mascheroni - 2.0 * log(4.0 * rho_used * t1) + + 2.0 * (Integrate_optimal (Integrand_11, args, 0, 0.0, t1, req_prec, 1.0e-30, 10000)).integ_est + - 2.0 * (Integrate_optimal (Integrand_12, args, 0, t1, tinf, req_prec, 1.0e-30, 10000)).integ_est + - (Integrate_optimal (Integrand_2, args, 0, 0.0, tinf, req_prec, 1.0e-30, 10000)).integ_est; + */ + return(answer); + } + + + /********************* TWO SPINONS ********************/ + + DP SF_2p (DP k, DP omega, I_table Itable) + { + // Careful ! This is S(k, omega) = S (k, w) |dw/domega| = 2 S(k, w) + + DP w = 2.0 * omega; // Rescale energies by factor 2 because of definitions of H_XXX (omega: S.S; w: 0.5 * sigma.sigma = 2 S.S) + + DP wu = twoPI * sin(0.5 * k); + DP wl = PI * fabs(sin(k)); + + // Factor of 2: return S(k, omega), not S(k, w) + // 0.25 factor: 1/4 * 2 * 1/2, where 1/4 comes from Bougourzi, 2 is the Jacobian |dw/domega| and 1/2 is S^{zz} = 1/2 * S^{+-} + return(w < wu && w > wl ? 2.0 * 0.5 * exp(-Itable.Return_val (acosh(sqrt((wu * wu - wl * wl)/(w * w - wl * wl)))/PI))/sqrt(wu * wu - w * w) : 0.0); + + } + + DP SF_2p (Vect_DP args, I_table Itable) + { + // Careful ! This is S(k, w) ! + + // This uses args[0] = k, args[1] = w. + + DP wu = twoPI * sin(0.5 * args[0]); + DP wl = PI * fabs(sin(args[0])); + + // 0.5 factor: 1 from Bougourzi, and 1/2 is S^{zz} = 1/2 * S^{+-} + return(args[1] < wu && args[1] > wl ? + 0.5 * exp(-Itable.Return_val (acosh(sqrt((wu * wu - wl * wl)/(args[1] * args[1] - wl * wl)))/PI))/sqrt(wu * wu - args[1] * args[1]) : 0.0); + } + + DP SF_2p_alt (Vect_DP args, I_table Itable) + { + // Careful ! This is S(k, w) ! + + // This uses args[0] = k, args[1] = alpha. + + DP wu = twoPI * sin(0.5 * args[0]); + DP wl = PI * fabs(sin(args[0])); + + //DP w = wu * cos(args[1]); + //DP factor = 1.0; + DP w = wl * cosh(args[1]); + + if (w >= wu || w <= wl) return(0.0); + + DP factor = sqrt((w * w - wl * wl)/(wu * wu - w * w)); + + // 0.5 factor: 1 from Bougourzi, and 1/2 is S^{zz} = 1/2 * S^{+-} + return(factor * 0.5 * exp(-Itable.Return_val (acosh(sqrt((wu * wu - wl * wl)/(w * w - wl * wl)))/PI))); + } + + DP SF_2p_w (Vect_DP args, I_table Itable) + { + return(args[1] * SF_2p (args, Itable)); + } + + DP SF_2p_w_alt (Vect_DP args, I_table Itable) + { + DP wu = twoPI * sin(0.5 * args[0]); + DP wl = PI * fabs(sin(args[0])); + + //DP w = wu * cos(args[1]); + //DP factor = 1.0; + DP w = wl * cosh(args[1]); + DP factor = sqrt((w * w - wl * wl)/(wu * wu - w * w)); + + // 0.5 factor: 1 from Bougourzi, and 1/2 is S^{zz} = 1/2 * S^{+-} + return(w * factor * 0.5 * exp(-Itable.Return_val (acosh(sqrt((wu * wu - wl * wl)/(w * w - wl * wl)))/PI))); + } + + DP SF_2p_intw (Vect_DP args, I_table Itable) + { + // This returns \int_0^2PI dw/2PI S(k, w) + + DP k = args[0]; + DP req_prec = args[1]; + int max_rec = int(args[2]); + + DP wu = twoPI * sin(0.5 * k); + DP wl = PI * fabs(sin(k)); + + Vect_DP args_to_SF_2p(2); + args_to_SF_2p[0] = k; + args_to_SF_2p[1] = 0.0; // this will be w + args_to_SF_2p[2] = JSC::max(1.0e-14, 0.01 * req_prec); + + return(Integrate_rec_using_table (SF_2p, args_to_SF_2p, 1, Itable, wl, wu, req_prec, max_rec)/twoPI); + } + + DP SF_2p_intw_alt (Vect_DP args, I_table Itable) + { + // This returns \int_0^2PI dw/2PI S(k, w) + + DP k = args[0]; + DP req_prec = args[1]; + int max_rec = int(args[2]); + + DP wu = twoPI * sin(0.5 * k); + DP wl = PI * fabs(sin(k)); + + Vect_DP args_to_SF_2p(2); + args_to_SF_2p[0] = k; + args_to_SF_2p[1] = 0.0; // this will be w + args_to_SF_2p[2] = JSC::max(1.0e-14, 0.01 * req_prec); + + //return(Integrate_rec_using_table (SF_2p_alt, args_to_SF_2p, 1, Itable, 0.0, acos(wl/wu), req_prec, max_rec)/twoPI); + return(Integrate_rec_using_table (SF_2p_alt, args_to_SF_2p, 1, Itable, 0.0, acosh(wu/wl), req_prec, max_rec)/twoPI); + } + + DP SF_2p_check_sumrule (DP req_prec, int max_rec, I_table Itable) + { + // It's better to use the ..._alt function below. + + Vect_DP args_to_SF_2p_intw (3); + + args_to_SF_2p_intw[0] = 0.0; // this will be k + args_to_SF_2p_intw[1] = JSC::max(1.0e-14, 0.01 * req_prec); + args_to_SF_2p_intw[2] = DP(max_rec); + + // Factor 2: int[0, 2PI] = 2 int[0, PI] + return(4.0 * 2.0 * Integrate_rec_using_table (SF_2p_intw, args_to_SF_2p_intw, 0, Itable, 0.0, PI, req_prec, max_rec)/twoPI); + // 4 : because full integral gives 1/4, return value here is sr fraction obtained. + } + + DP SF_2p_check_sumrule_alt (DP req_prec, int max_rec, I_table Itable) + { + // This is the preferred version. + + Vect_DP args_to_SF_2p_intw (3); + + args_to_SF_2p_intw[0] = 0.0; // this will be k + args_to_SF_2p_intw[1] = JSC::max(1.0e-14, 0.01 * req_prec); + args_to_SF_2p_intw[2] = DP(max_rec); + + // Factor 2: int[0, 2PI] = 2 int[0, PI] + return(4.0 * 2.0 * Integrate_rec_using_table (SF_2p_intw_alt, args_to_SF_2p_intw, 0, Itable, 0.0, PI, req_prec, max_rec)/twoPI); + // 4 : because full integral gives 1/4, return value here is sr fraction obtained. + } + + DP Fixed_k_sumrule_w (DP k) + { + // This is K_1 (k) = \int dw/2PI w S(k, w) = 2 K_1^{KarbachPRB55} (k) = 4 E_G (1-cosk)/3N with E_G = -N(ln2 - 1/4). + return(4.0 * (log(2.0) - 0.25) * (1.0 - cos(k))/3.0); + } + + DP Fixed_k_sumrule_omega (DP k) + { + return(0.5 * Fixed_k_sumrule_w(k)); + } + + DP SF_2p_check_fixed_k_sumrule (DP k, DP req_prec, int max_rec, I_table Itable) + { + // It's better to use the ..._alt function below. + + DP wu = twoPI * sin(0.5 * k); + DP wl = PI * fabs(sin(k)); + + Vect_DP args_to_SF_2p(2); + args_to_SF_2p[0] = k; + args_to_SF_2p[1] = 0.0; // this will be w + args_to_SF_2p[2] = JSC::max(1.0e-14, 0.01 * req_prec); + + return((Integrate_rec_using_table (SF_2p_w, args_to_SF_2p, 1, Itable, wl, wu, req_prec, max_rec)/twoPI)/Fixed_k_sumrule_w(k)); + } + + DP SF_2p_check_fixed_k_sumrule_alt (DP k, DP req_prec, int max_rec, I_table Itable) + { + // This is the preferred version. + + DP wu = twoPI * sin(0.5 * k); + DP wl = PI * fabs(sin(k)); + + Vect_DP args_to_SF_2p(2); + args_to_SF_2p[0] = k; + args_to_SF_2p[1] = 0.0; // this will be alpha + args_to_SF_2p[2] = JSC::max(1.0e-14, 0.01 * req_prec); + + return((Integrate_rec_using_table (SF_2p_w_alt, args_to_SF_2p, 1, Itable, 0.0, acosh(wu/wl), req_prec, max_rec)/twoPI)/Fixed_k_sumrule_w(k)); + } + + DP SF_2p_check_fixed_k_sumrule_opt (DP k, DP req_prec, int Npts, I_table Itable) + { + // This is the preferred version. + + DP wu = twoPI * sin(0.5 * k); + DP wl = PI * fabs(sin(k)); + + Vect_DP args_to_SF_2p(2); + args_to_SF_2p[0] = k; + args_to_SF_2p[1] = 0.0; // this will be alpha + args_to_SF_2p[2] = JSC::max(1.0e-14, 0.01 * req_prec); + + return(((Integrate_optimal_using_table (SF_2p_w_alt, args_to_SF_2p, 1, Itable, 0.0, acosh(wu/wl), req_prec, 1.0e-32, Npts)).integ_est/twoPI)/Fixed_k_sumrule_w(k)); + } + + + + /********************** FOUR SPINONS **********************/ + + + DP Sum_norm_gl (Vect_DP rho, DP req_prec) + { + complex g[4]; + for (int l = 0; l < 4; ++l) g[l] = 0.0; + + //DP prefactor = 256.0 * pow(PI, 14.0); + //DP prefactor = 1.0; // All factors taken into account later + + complex Plm[4]; + complex Pm[4]; + + DP den = 0.0; + for (int j = 0; j < 4; ++j) { + Pm[j] = cosh(twoPI * rho[j]); + den = 1.0; + for (int i = 0; i < 4; ++i) if (i != j) den *= sinh(PI * (rho[j] - rho[i])); + Pm[j] /= den; + } + + complex irhoj0[4]; + complex irhoj1[4]; + complex irhoj2[4]; + complex irhoj3[4]; + + for (int j = 0; j < 4; ++j) { + irhoj0[j] = II * (rho[j] - rho[0]); + irhoj1[j] = II * (rho[j] - rho[1]); + irhoj2[j] = II * (rho[j] - rho[2]); + irhoj3[j] = II * (rho[j] - rho[3]); + } + + // Do m = 0 terms: + /* + Pm[0] *= Gamma_min_0p5 * exp(ln_Gamma(-0.5 + irhoj1[0]) - ln_Gamma(1.0 + irhoj1[0]) + + ln_Gamma(-0.5 + irhoj2[0]) - ln_Gamma(1.0 + irhoj2[0]) + + ln_Gamma(-0.5 + irhoj3[0]) - ln_Gamma(1.0 + irhoj3[0])); + Pm[1] *= Gamma_min_0p5 * exp(ln_Gamma(-0.5 + irhoj0[1]) - ln_Gamma(1.0 + irhoj0[1]) + + ln_Gamma(-0.5 + irhoj2[1]) - ln_Gamma(1.0 + irhoj2[1]) + + ln_Gamma(-0.5 + irhoj3[1]) - ln_Gamma(1.0 + irhoj3[1])); + Pm[2] *= Gamma_min_0p5 * exp(ln_Gamma(-0.5 + irhoj0[2]) - ln_Gamma(1.0 + irhoj0[2]) + + ln_Gamma(-0.5 + irhoj1[2]) - ln_Gamma(1.0 + irhoj1[2]) + + ln_Gamma(-0.5 + irhoj3[2]) - ln_Gamma(1.0 + irhoj3[2])); + Pm[3] *= Gamma_min_0p5 * exp(ln_Gamma(-0.5 + irhoj0[3]) - ln_Gamma(1.0 + irhoj0[3]) + + ln_Gamma(-0.5 + irhoj1[3]) - ln_Gamma(1.0 + irhoj1[3]) + + ln_Gamma(-0.5 + irhoj2[3]) - ln_Gamma(1.0 + irhoj2[3])); + */ + for (int j = 0; j < 4; ++j) { + /* + Pm[j] *= exp(ln_Gamma(-0.5 + irhoj0[j]) - ln_Gamma(1.0 + irhoj0[j]) + + ln_Gamma(-0.5 + irhoj1[j]) - ln_Gamma(1.0 + irhoj1[j]) + + ln_Gamma(-0.5 + irhoj2[j]) - ln_Gamma(1.0 + irhoj2[j]) + + ln_Gamma(-0.5 + irhoj3[j]) - ln_Gamma(1.0 + irhoj3[j])); + */ + // Calling only Gamma (z) for Re(z) >= 0.5, in view of Lanczos method: + Pm[j] *= exp(ln_Gamma(0.5 + irhoj0[j]) - ln_Gamma(1.0 + irhoj0[j]) + + ln_Gamma(0.5 + irhoj1[j]) - ln_Gamma(1.0 + irhoj1[j]) + + ln_Gamma(0.5 + irhoj2[j]) - ln_Gamma(1.0 + irhoj2[j]) + + ln_Gamma(0.5 + irhoj3[j]) - ln_Gamma(1.0 + irhoj3[j])) + /((-0.5 + irhoj0[j]) * (-0.5 + irhoj1[j]) * (-0.5 + irhoj2[j]) * (-0.5 + irhoj3[j])); + + for (int l = 0; l < 4; ++l) { + + Plm[j] = 1.0; + for (int i = 0; i < 4; ++i) if (i != l) Plm[j] *= ((l > i ? -0.5 : 0.0) + II * (rho[j] - rho[i])); + + if (j <= l) g[l] += Plm[j] * Pm[j]; // otherwise no m = 0 term + + //cout << "j = " << j << "\tl = " << l << "\tPlm[j] = " << Plm[j] << "\tPm[j] = " << Pm[j] << "\tprod = " << Plm[j] * Pm[j] << endl; + } + } + + /* + for (int j = 0; j < 4; ++j) { + if (j == 0) g[0] += irhoj1[j] * irhoj2[j] * irhoj3[j] * Pm[j]; + if (j <= 1) g[1] += (-0.5 + irhoj0[j]) * irhoj2[j] * irhoj3[j] * Pm[j]; + if (j <= 2) g[2] += (-0.5 + irhoj0[j]) * (-0.5 + irhoj1[j]) * irhoj3[j] * Pm[j]; + g[3] += (-0.5 + irhoj0[j]) * (-0.5 + irhoj1[j]) * (-0.5 + irhoj2[j]) * Pm[j]; + } + */ + + DP sum_norm_gl = norm(g[0]) + norm(g[1]) + norm(g[2]) + norm(g[3]); + DP old_sum_norm_gl = sum_norm_gl; + + //cout << "|g1|^2 = " << prefactor * norm(g[0]) << "\t2 " << prefactor * norm(g[1]) << "\t3 " << prefactor * norm(g[2]) << "\t4 " << prefactor * norm(g[3]) << endl; + + // Do m = 1, 2, ... terms: + + int m = 1; + int m_to_reach = 1; + + do { + + old_sum_norm_gl = sum_norm_gl; + + // We increase m by ten steps before checking sum_norm + + m_to_reach = m + 10; + + do { + + for (int j = 0; j < 4; ++j) { + + Pm[j] *= (m - 1.5 + irhoj0[j]) * (m - 1.5 + irhoj1[j]) * (m - 1.5 + irhoj2[j]) * (m - 1.5 + irhoj3[j]) + / ((DP(m) + irhoj0[j]) * (DP(m) + irhoj1[j]) * (DP(m) + irhoj2[j]) * (DP(m) + irhoj3[j])); + + //for (int l = 0; l < 4; ++l) { + + //Plm[j] = 1.0; + //for (int i = 0; i < 4; ++i) if (i != l) Plm[j] *= m - (l > i ? 0.5 : 0.0) + II * (rho[j] - rho[i]); + + //g[l] += Plm[j] * Pm[j]; + //} + + // FASTER: unwrap l, i loops + // l = 0: + //Plm[j] = (DP(m) + II * (rho[j] - rho[1])) * (DP(m) + II * (rho[j] - rho[2])) * (DP(m) + II * (rho[j] - rho[3])); + //Plm[j] = (DP(m) + irhoj1[j]) * (DP(m) + irhoj2[j]) * (DP(m) + irhoj3[j]); + //g[0] += Plm[j] * Pm[j]; + g[0] += (DP(m) + irhoj1[j]) * (DP(m) + irhoj2[j]) * (DP(m) + irhoj3[j]) * Pm[j]; + // l = 1; + //Plm[j] = (m - 0.5 + II * (rho[j] - rho[0])) * (DP(m) + II * (rho[j] - rho[2])) * (DP(m) + II * (rho[j] - rho[3])); + //Plm[j] = (m - 0.5 + irhoj0[j]) * (DP(m) + irhoj2[j]) * (DP(m) + irhoj3[j]); + //g[1] += Plm[j] * Pm[j]; + g[1] += (m - 0.5 + irhoj0[j]) * (DP(m) + irhoj2[j]) * (DP(m) + irhoj3[j]) * Pm[j]; + // l = 2; + //Plm[j] = (m - 0.5 + II * (rho[j] - rho[0])) * (m - 0.5 + II * (rho[j] - rho[1])) * (DP(m) + II * (rho[j] - rho[3])); + //Plm[j] = (m - 0.5 + irhoj0[j]) * (m - 0.5 + irhoj1[j]) * (DP(m) + irhoj3[j]); + //g[2] += Plm[j] * Pm[j]; + g[2] += (m - 0.5 + irhoj0[j]) * (m - 0.5 + irhoj1[j]) * (DP(m) + irhoj3[j]) * Pm[j]; + // l = 3; + //Plm[j] = (m - 0.5 + II * (rho[j] - rho[0])) * (m - 0.5 + II * (rho[j] - rho[1])) * (m - 0.5 + II * (rho[j] - rho[2])); + //Plm[j] = (m - 0.5 + irhoj0[j]) * (m - 0.5 + irhoj1[j]) * (m - 0.5 + irhoj2[j]); + //g[3] += Plm[j] * Pm[j]; + g[3] += (m - 0.5 + irhoj0[j]) * (m - 0.5 + irhoj1[j]) * (m - 0.5 + irhoj2[j]) * Pm[j]; + } + + /* + // Also unwrap j loop: + Pm[0] *= (m - 1.5 + irhoj0[0]) * (m - 1.5 + irhoj1[0]) * (m - 1.5 + irhoj2[0]) * (m - 1.5 + irhoj3[0]) + / ((DP(m) + irhoj0[0]) * (DP(m) + irhoj1[0]) * (DP(m) + irhoj2[0]) * (DP(m) + irhoj3[0])); + Pm[1] *= (m - 1.5 + irhoj0[1]) * (m - 1.5 + irhoj1[1]) * (m - 1.5 + irhoj2[1]) * (m - 1.5 + irhoj3[1]) + / ((DP(m) + irhoj0[1]) * (DP(m) + irhoj1[1]) * (DP(m) + irhoj2[1]) * (DP(m) + irhoj3[1])); + Pm[2] *= (m - 1.5 + irhoj0[2]) * (m - 1.5 + irhoj1[2]) * (m - 1.5 + irhoj2[2]) * (m - 1.5 + irhoj3[2]) + / ((DP(m) + irhoj0[2]) * (DP(m) + irhoj1[2]) * (DP(m) + irhoj2[2]) * (DP(m) + irhoj3[2])); + Pm[3] *= (m - 1.5 + irhoj0[3]) * (m - 1.5 + irhoj1[3]) * (m - 1.5 + irhoj2[3]) * (m - 1.5 + irhoj3[3]) + / ((DP(m) + irhoj0[3]) * (DP(m) + irhoj1[3]) * (DP(m) + irhoj2[3]) * (DP(m) + irhoj3[3])); + + g[0] += ((DP(m) + irhoj1[0]) * (DP(m) + irhoj2[0]) * (DP(m) + irhoj3[0])) * Pm[0] + + ((DP(m)) * (DP(m) + irhoj2[1]) * (DP(m) + irhoj3[1])) * Pm[1] + + ((DP(m) + irhoj1[2]) * (DP(m)) * (DP(m) + irhoj3[0])) * Pm[2] + + ((DP(m) + irhoj1[3]) * (DP(m) + irhoj2[3]) * (DP(m))) * Pm[3]; + + g[1] += ((m - 0.5) * (DP(m) + irhoj2[0]) * (DP(m) + irhoj3[0])) * Pm[0] + + ((m - 0.5 + irhoj0[1]) * (DP(m) + irhoj2[1]) * (DP(m) + irhoj3[1])) * Pm[1] + + ((m - 0.5 + irhoj0[2]) * (DP(m)) * (DP(m) + irhoj3[2])) * Pm[2] + + ((m - 0.5 + irhoj0[3]) * (DP(m) + irhoj2[3]) * (DP(m))) * Pm[3]; + + g[2] += ((m - 0.5) * (m - 0.5 + irhoj1[0]) * (DP(m) + irhoj3[0])) * Pm[0] + + ((m - 0.5 + irhoj0[1]) * (m - 0.5) * (DP(m) + irhoj3[1])) * Pm[1] + + ((m - 0.5 + irhoj0[2]) * (m - 0.5 + irhoj1[2]) * (DP(m) + irhoj3[2])) * Pm[2] + + ((m - 0.5 + irhoj0[3]) * (m - 0.5 + irhoj1[3]) * (DP(m))) * Pm[3]; + + g[3] += ((m - 0.5) * (m - 0.5 + irhoj1[0]) * (m - 0.5 + irhoj2[0])) * Pm[0] + + ((m - 0.5 + irhoj0[1]) * (m - 0.5) * (m - 0.5 + irhoj2[1])) * Pm[1] + + ((m - 0.5 + irhoj0[2]) * (m - 0.5 + irhoj1[2]) * (m - 0.5)) * Pm[2] + + ((m - 0.5 + irhoj0[3]) * (m - 0.5 + irhoj1[3]) * (m - 0.5 + irhoj2[3])) * Pm[3]; + */ + m++; + + } while (m < m_to_reach); + + sum_norm_gl = norm(g[0]) + norm(g[1]) + norm(g[2]) + norm(g[3]); + + } while (m < 10 || sum_norm_gl/old_sum_norm_gl - 1.0 > req_prec && m < 100000); + + //cout << "g converged using " << m << " terms." << endl; + + //cout << "|g1|^2 = " << prefactor * norm(g[0]) << "\t2 " << prefactor * norm(g[1]) << "\t3 " << prefactor * norm(g[2]) << "\t4 " << prefactor * norm(g[3]) << endl; + + //return(prefactor * (norm(g[0]) + norm(g[1]) + norm(g[2]) + norm(g[3]))); + return(norm(g[0]) + norm(g[1]) + norm(g[2]) + norm(g[3])); + + } + + DP Compute_C4 (DP req_prec) + { + Vect_DP args(2); + + DP answer = exp(-8.0 * real(ln_Gamma (0.25)) - 9.0 * log(2.0) + 8.0 * Integrate_rec (Integrand_A, args, 0, 0.0, 50.0, req_prec, 16))/3.0; + //DP answer = exp(-8.0 * real(ln_Gamma (0.25)) - 21.0 * log(2.0) - 14.0 * log(PI) + 8.0 * Integrate_rec (Integrand_A, args, 0, 0.0, 100.0, req_prec, 16))/3.0; + + //cout << "|A|^2 = " << exp(-2.0 * Integrate_rec (Integrand_A, args, 0, 0.0, 100.0, req_prec)) + // << "\t Gamma (1/4) = " << exp ((ln_Gamma(0.25))) << endl; + //cout << "NPB: " << exp(-8.0 * real(ln_Gamma (0.25)) - 9.0 * log(2.0) + 8.0 * Integrate_rec (Integrand_A, args, 0, 0.0, 50.0, req_prec))/3.0 << endl; + //cout << "c-m: " << exp(-8.0 * real(ln_Gamma (0.25)) - 21.0 * log(2.0) - 14.0 * log(PI) + 8.0 * Integrate_rec (Integrand_A, args, 0, 0.0, 50.0, req_prec))/3.0 << endl; + + return(answer); + } + + DP SF_contrib (Vect_DP p, DP req_prec, I_table Itable) + { + Vect_DP rho(4); + DP W, Wu, sum_I_integrals, sum_norm_g; + + for (int i = 0; i < 4; ++i) rho[i] = asinh(1.0/tan(p[i]))/twoPI; + + Wu = twoPI* fabs(sin(0.5*(p[0] + p[1]))); + W = -PI* (sin(p[0]) + sin(p[1])); + + sum_I_integrals = 0.0; + for (int i1 = 0; i1 < 3; ++i1) for (int i2 = i1+1; i2 < 4; ++i2) { + //cout << "rho_ij = " << rho[i1] - rho[i2] << "\tI(rho_ij) = " << Itable.Return_val (rho[i1] - rho[i2]) << "\t" << I_integral (rho[i1] - rho[i2], req_prec) << endl; + //sum_I_integrals += I_integral(rho[i1] - rho[i2], req_prec); + sum_I_integrals += Itable.Return_val (rho[i1] - rho[i2]); + } + //cout << "sum_I_integrals = " << sum_I_integrals << "\texp(-sum) = " << exp(-sum_I_integrals) << endl; + + //sum_norm_g = 0.0; + //for (int i = 0; i < 4; ++i) sum_norm_g += norm(g(i, rho)); + sum_norm_g = Sum_norm_gl (rho, req_prec); + + //cout << "sum_norm_g = " << sum_norm_g << "\t sqrt() = " << sqrt(Wu * Wu - W * W) << endl; + + return(exp(-sum_I_integrals) * sum_norm_g/sqrt(Wu * Wu - W * W)); + } + + DP J_fn (Vect_DP p, DP req_prec, I_table Itable) + { + Vect_DP rho(4); + DP sum_I_integrals, sum_norm_g; + + for (int i = 0; i < 4; ++i) rho[i] = asinh(1.0/tan(p[i]))/twoPI; + + sum_I_integrals = 0.0; + for (int i1 = 0; i1 < 3; ++i1) for (int i2 = i1+1; i2 < 4; ++i2) { + if (fabs(rho[i1] - rho[i2]) < 1.0e-10 || fabs(rho[i1] - rho[i2]) >= 1000.0) return(0.0); // safety here + sum_I_integrals += Itable.Return_val (rho[i1] - rho[i2]); + } + + sum_norm_g = Sum_norm_gl (rho, req_prec); + + return(exp(-sum_I_integrals) * sum_norm_g); + } + + bool Set_p_given_kwKW (DP k, DP w, DP K, DP W, Vect_DP& p) + { + // Returns false if any of the p_i are out of -PI, 0 + + DP argacos1, argacos2; + + if (fabs(argacos1 = W/(twoPI * sin(0.5*K))) > 1.0 || fabs(argacos2 = (w - W)/(twoPI * sin (0.5 * fabs(k - K)))) > 1.0) return(false); + + DP acos1 = acos(argacos1); + DP acos2 = acos(argacos2); + + p[0] = -0.5 * K + acos1; + p[1] = -0.5 * K - acos1; + + if (K <= k) { + p[2] = 0.5 * (K-k) + acos2; + p[3] = 0.5 * (K-k) - acos2; + } + + else { + p[2] = 0.5 * (K-k) - PI + acos2; + p[3] = 0.5 * (K-k) - PI - acos2; + } + + for (int i = 0; i < 4; ++i) if (p[i] < -PI || p[i] > 0.0) return(false); // { cout << k << "\t" << w << "\t" << K << "\t" << W << "\t" << p; JSCerror("p out of bounds"); } + + return(true); + } + + DP G_fn (Vect_DP args_to_G, I_table Itable) + { + Vect_DP p(4); + + if (!Set_p_given_kwKW (args_to_G[0], args_to_G[1], args_to_G[2], args_to_G[3], p)) return(0.0); + + DP answer = Jacobian_p3p4_KW (args_to_G[0], args_to_G[1], args_to_G[2], args_to_G[3]) * SF_contrib (p, args_to_G[4], Itable); + + // cout << "kwKW = " << args_to_G[0] << " " << args_to_G[1] << " " << args_to_G[2] << " " << args_to_G[3] << "\tp = " << p << "Jac = " << Jacobian_p3p4_KW (args_to_G[0], args_to_G[1], args_to_G[2], args_to_G[3]) << "\tG = " << answer << endl; + + //return(Jacobian_p3p4_KW (args_to_G[0], args_to_G[1], args_to_G[2], args_to_G[3]) * SF_contrib (p, 0.01 * args_to_G[4], Itable)); + return(answer); + } + + DP G1_fn (Vect_DP args_to_G, I_table Itable) + { + Vect_DP p(4); + + DP W = twoPI * sin(0.5*args_to_G[2]) * cos(args_to_G[3]); + + if (!Set_p_given_kwKW (args_to_G[0], args_to_G[1], args_to_G[2], W, p)) return(0.0); + + return(J_fn (p, args_to_G[4], Itable)/sqrt(pow(twoPI * sin(0.5*(args_to_G[0] - args_to_G[2])), 2.0) + - pow(args_to_G[1] - W, 2.0))); + } + + DP G2_fn (Vect_DP args_to_G, I_table Itable) + { + Vect_DP p(4); + + DP W = args_to_G[1] - twoPI * fabs(sin(0.5*(args_to_G[0] - args_to_G[2]))) * cos(args_to_G[3]); + + if (!Set_p_given_kwKW (args_to_G[0], args_to_G[1], args_to_G[2], W, p)) return(0.0); + + return(J_fn (p, args_to_G[4], Itable)/sqrt(pow(twoPI * sin(0.5*args_to_G[2]), 2.0) - W * W)); + } + + DP G1_fn_mid (Vect_DP args_to_G, I_table Itable) + { + // Called by H_fn_mid. + + // For the lower half of W interval + + // Translation of arguments to G: + // args_to_G[0] = k; + // args_to_G[1] = w; + // args_to_G[2] = K; + // args_to_G[3] = alpha; + // args_to_G[4] = req_prec; + // args_to_G[5] = max_rec; + // args_to_G[6] = Wmid; + // args_to_G[7] = Wmax_used; + // args_to_G[8] = Wu_sq; + // args_to_G[9] = Wut_sq; + + Vect_DP p(4); + + DP W = args_to_G[6] * (args_to_G[3] > 1.0e-4 ? 1.0 - cos(args_to_G[3]) : 2.0 * pow(sin(0.5 * args_to_G[3]), 2)); + // W = Wmid (1 - cos(alpha)), ensure some precision if alpha small + + if (!Set_p_given_kwKW (args_to_G[0], args_to_G[1], args_to_G[2], W, p)) return(0.0); + + DP answer = J_fn (p, args_to_G[4], Itable) * sqrt(W * (2.0 * args_to_G[6] - W) + /((args_to_G[8] - W * W) * (args_to_G[9] - pow(args_to_G[1] - W, 2.0)))); + + if (is_nan(answer)) { + cerr << setprecision(10) << "args_to_G1_fn_mid = " << args_to_G << "G1 = " << answer << "\tPut to zero..." << endl; + //JSCerror("non !"); + answer = 0.0; + } + return(answer); + } + + DP G2_fn_mid (Vect_DP args_to_G, I_table Itable) + { + // Called by H_fn_mid. + + // For the upper half of W interval + + // See above for translation of arguments to G. + + Vect_DP p(4); + + DP W = args_to_G[7] * cos(args_to_G[3]); // W = Wmax cos(alpha) + + if (!Set_p_given_kwKW (args_to_G[0], args_to_G[1], args_to_G[2], W, p)) return(0.0); + + DP answer = J_fn (p, args_to_G[4], Itable) + //* sqrt((args_to_G[7] * args_to_G[7] - W * W)/((args_to_G[8] - W * W) * (args_to_G[9] - pow(args_to_G[1] - W, 2.0)))); + * args_to_G[7] * sin(args_to_G[3]) /sqrt((args_to_G[8] - W * W) * (args_to_G[9] - pow(args_to_G[1] - W, 2.0))); + + if (is_nan(answer)) { + cerr << setprecision(10) << "args_to_G2_fn_mid = " << args_to_G << "G2 = " << answer << endl; + cerr << W << "\t" << (args_to_G[7] * args_to_G[7] - W * W) << "\t" << (args_to_G[8] - W * W) << "\t" << (args_to_G[9] - pow(args_to_G[1] - W, 2.0)) << endl; + //JSCerror("non !"); + answer = 0.0; + } + return(answer); + } + + + DP G_fn_alt (Vect_DP args_to_G, I_table Itable) + { + Vect_DP p(4); + + DP Wmin = args_to_G[4]; + DP W = Wmin * cosh(args_to_G[3]); + + if (!Set_p_given_kwKW (args_to_G[0], args_to_G[1], args_to_G[2], W, p)) return(0.0); + + DP Wu1 = args_to_G[6]; + DP Wu2 = args_to_G[7]; + + return(J_fn (p, args_to_G[8], Itable) * sqrt((W * W - Wmin * Wmin)/((Wu1 * Wu1 - W * W) * (Wu2 * Wu2 - (args_to_G[1] - W) * (args_to_G[1] - W))))); + } + + DP H_fn (Vect_DP args_to_H, I_table Itable) + { + // Translate arguments to more readable form + DP k = args_to_H[0]; + DP w = args_to_H[1]; + DP K = args_to_H[2]; + DP req_prec = JSC::max(1.0e-14, args_to_H[3]); + int max_rec = int(args_to_H[4]); + + Vect_DP args_to_G(6); + args_to_G[0] = k; + args_to_G[1] = w; + args_to_G[2] = K; + args_to_G[3] = 0.0; + args_to_G[4] = JSC::max(1.0e-14, 0.01 * req_prec); + args_to_G[5] = DP(max_rec); + + DP Wmin_used = Wmin (k, w, K); + DP Wmax_used = Wmax (k, w, K); + + return(Wmax_used > Wmin_used ? Integrate_rec_using_table (G_fn, args_to_G, 3, Itable, Wmin_used, Wmax_used, req_prec, max_rec) : 0.0); + //return(Riemann_Integrate_rec_using_table (G_fn, args_to_G, 3, Itable, 0.0, 10.0, 500)); + } + + DP H2_fn (Vect_DP args_to_H, I_table Itable) + { + // Translate arguments to more readable form + DP k = args_to_H[0]; + DP w = args_to_H[1]; + DP K = args_to_H[2]; + DP req_prec = JSC::max(1.0e-14, args_to_H[3]); + int max_rec = int(args_to_H[4]); + + DP Wmin_used = Wmin (k, w, K); + DP Wmax_used = Wmax (k, w, K); + + if (Wmax_used <= Wmin_used) return(0.0); + + Vect_DP args_to_G(6); + args_to_G[0] = k; + args_to_G[1] = w; + args_to_G[2] = K; + args_to_G[3] = 0.0; // this will be the alpha parameter + args_to_G[4] = JSC::max(1.0e-14, 0.01 * req_prec); + args_to_G[5] = DP(max_rec); + + DP Wmid = 0.5 * (Wmin_used + Wmax_used); + + DP answer = 0.0; + + DP alpha_L1 = acos(JSC::min(1.0, Wmax_used/(twoPI * sin(0.5*K)))); // to prevent nan + DP alpha_U1 = acos(Wmid/(twoPI * sin(0.5*K))); + DP alpha_L2 = acos(JSC::min(1.0, (w - Wmin_used)/(twoPI * fabs(sin(0.5*(k - K)))))); + DP alpha_U2 = acos((w - Wmid)/(twoPI * fabs(sin(0.5*(k - K))))); + + answer += Integrate_rec_using_table (G1_fn, args_to_G, 3, Itable, alpha_L1, alpha_U1, req_prec, max_rec); + answer += Integrate_rec_using_table (G2_fn, args_to_G, 3, Itable, alpha_L2, alpha_U2, req_prec, max_rec); + + return(answer); + } + + DP H_fn_mid (Vect_DP args_to_H, I_table Itable) + { + // For W in [Wmin, Wmid] we use the parametrization W = Wmid sin(alpha), alpha in [0, PI/2] + // such that dW = Wmid cos(alpha) dalpha is approx. 0 around alpha = PI/2 (W = 0). + // For W in [Wmid, Wmax] we use W = Wmax cos(alpha), alpha in [0, acos(Wmid/Wmax)] + // such that dW = -Wmax sin(alpha) dalpha is approx 0 around alpha = 0 (W = Wmax). + + // Translation of args_to_H: + // args_to_H[0] = k; + // args_to_H[1] = w; + // args_to_H[2] = K; <------ integrated over, so newly set here + // args_to_H[3] = req_prec; + // args_to_H[4] = max_rec; + + DP Wmin_used = Wmin (args_to_H[0], args_to_H[1], args_to_H[2]); + DP Wmax_used = Wmax (args_to_H[0], args_to_H[1], args_to_H[2]); + + DP Wu_sq = pow(twoPI * sin(0.5 * args_to_H[2]), 2.0); + DP Wut_sq = pow(twoPI * sin(0.5 * (args_to_H[0] - args_to_H[2])), 2.0); + + if (Wmax_used <= Wmin_used) return(0.0); + + DP Wmid = 0.5 * (Wmin_used + Wmax_used); + + Vect_DP args_to_G(10); + args_to_G[0] = args_to_H[0]; + args_to_G[1] = args_to_H[1]; + args_to_G[2] = args_to_H[2]; + args_to_G[3] = 0.0; // this will be the alpha parameter + args_to_G[4] = JSC::max(1.0e-14, 0.01 * args_to_H[3]); + args_to_G[5] = args_to_H[4]; + args_to_G[6] = Wmid; + args_to_G[7] = Wmax_used; + args_to_G[8] = Wu_sq; + args_to_G[9] = Wut_sq; + + DP answer = 0.0; + + DP alpha_L1 = 0.0; + DP alpha_U1 = 0.5 * PI; + DP alpha_L2 = 0.0; + DP alpha_U2 = acos(Wmid/Wmax_used); + + answer += Integrate_rec_using_table (G1_fn_mid, args_to_G, 3, Itable, alpha_L1, alpha_U1, args_to_H[3], int(args_to_H[4])); + answer += Integrate_rec_using_table (G2_fn_mid, args_to_G, 3, Itable, alpha_L2, alpha_U2, args_to_H[3], int(args_to_H[4])); + + return(answer); + } + + DP H_fn_mid_opt (Vect_DP args_to_H, I_table Itable) + { + // For W in [Wmin, Wmid] we use the parametrization W = Wmid sin(alpha), alpha in [0, PI/2] + // such that dW = Wmid cos(alpha) dalpha is approx. 0 around alpha = PI/2 (W = 0). + // For W in [Wmid, Wmax] we use W = Wmax cos(alpha), alpha in [0, acos(Wmid/Wmax)] + // such that dW = -Wmax sin(alpha) dalpha is approx 0 around alpha = 0 (W = Wmax). + + // Translation of args_to_H: + // args_to_H[0] = k; + // args_to_H[1] = w; + // args_to_H[2] = K; <------ integrated over, so newly set here + // args_to_H[3] = req_prec; + // args_to_H[4] = Npts; + + DP Wmin_used = Wmin (args_to_H[0], args_to_H[1], args_to_H[2]); + DP Wmax_used = Wmax (args_to_H[0], args_to_H[1], args_to_H[2]); + + DP Wu_sq = pow(twoPI * sin(0.5 * args_to_H[2]), 2.0); + DP Wut_sq = pow(twoPI * sin(0.5 * (args_to_H[0] - args_to_H[2])), 2.0); + + if (Wmax_used <= Wmin_used) return(0.0); + + DP Wmid = 0.5 * (Wmin_used + Wmax_used); + + Vect_DP args_to_G(10); + args_to_G[0] = args_to_H[0]; + args_to_G[1] = args_to_H[1]; + args_to_G[2] = args_to_H[2]; + args_to_G[3] = 0.0; // this will be the alpha parameter + args_to_G[4] = JSC::max(1.0e-14, 0.01 * args_to_H[3]); + args_to_G[5] = args_to_H[4]; + args_to_G[6] = Wmid; + args_to_G[7] = Wmax_used; + args_to_G[8] = Wu_sq; + args_to_G[9] = Wut_sq; + + DP answer = 0.0; + + DP alpha_L1 = 0.0; + DP alpha_U1 = 0.5 * PI; + DP alpha_L2 = 0.0; + DP alpha_U2 = acos(Wmid/Wmax_used); + + answer += (Integrate_optimal_using_table (G1_fn_mid, args_to_G, 3, Itable, alpha_L1, alpha_U1, args_to_H[3], 1.0e-32, int(args_to_H[4]))).integ_est; + answer += (Integrate_optimal_using_table (G2_fn_mid, args_to_G, 3, Itable, alpha_L2, alpha_U2, args_to_H[3], 1.0e-32, int(args_to_H[4]))).integ_est; + + return(answer); + } + + DP H_fn_alt (Vect_DP args_to_H, I_table Itable) + { + // Translate arguments to more readable form + DP k = args_to_H[0]; + DP w = args_to_H[1]; + DP K = args_to_H[2]; + DP req_prec = JSC::max(1.0e-14, args_to_H[3]); + int max_rec = int(args_to_H[4]); + + DP Wmin_used = Wmin (k, w, K); + DP Wmax_used = Wmax (k, w, K); + + if (Wmax_used <= Wmin_used) return(0.0); + + Vect_DP args_to_G(10); + args_to_G[0] = k; + args_to_G[1] = w; + args_to_G[2] = K; + args_to_G[3] = 0.0; // this will become alpha + args_to_G[4] = Wmin_used; + args_to_G[5] = Wmax_used; + args_to_G[6] = twoPI * fabs(sin(0.5*args_to_G[2])); // Wu1 + args_to_G[7] = twoPI * fabs(sin(0.5*(args_to_G[0] - args_to_G[2]))); // Wu2 + args_to_G[8] = JSC::max(1.0e-14, 0.01 * req_prec); + args_to_G[9] = DP(max_rec); + + return(Integrate_rec_using_table (G_fn_alt, args_to_G, 3, Itable, 0.0, acosh(Wmax_used/Wmin_used), req_prec, max_rec)); + //return(Riemann_Integrate_rec_using_table (G_fn, args_to_G, 3, Itable, 0.0, 10.0, 500)); + } + + //DP SF_4p_kwKW (DP k, DP omega, DP req_prec, int max_rec, I_table Itable) + DP SF_4p_kwKW (Vect_DP args, I_table Itable) + { + // Translate: + // args[0] = k; + // args[1] = omega; + // args[2] = req_prec; + // args[3] = max_rec; + DP k = args[0]; + DP omega = args[1]; + DP req_prec = args[2]; + int max_rec = int(args[3]); + + Vect_DP args_to_H(5); + args_to_H[0] = k; // shift of PI in Bougourzi: because they do FM case. + // We want AFM, so SF_4p (k, omega) is correctly obtained directly from the RHS of their formula. + DP w = 2.0 * omega; // Rescale energies by factor 2 because of definitions of H_XXX (omega: S.S; w: 0.5 * sigma.sigma = 2 S.S) + args_to_H[1] = w; + args_to_H[2] = 0.0; // this is K + args_to_H[3] = JSC::max(1.0e-14, 0.01 * req_prec); + args_to_H[4] = DP(max_rec); + + if (w > wmax_4p(k) || w < wmin_4p(k)) { + //cout << "w out of bounds in SF_4p: " << "wmin = " << PI * sin(k) << " wmax = " << 4.0 * PI * sin(0.25 * k) << " w = " << w << endl; + return(0.0); + } + + DP prefactor = 2.0 * 0.5 * 4.0 * Compute_C4 (req_prec); // 4 comes from using p1 > p2 & p3 > p4 instead of whole interval. + // 2 from Jacobian |dw/domega| + // 0.5 from S^{zz} = S^{pm}/2 + //DP prefactor = 4.0; + + // Define the K integral domain + Domain Kdomain; + + // First, the inclusions: + if (w <= twoPI * sin(0.5*k)) Kdomain.Include (0.0, twoPI); + + else { + if (w < 4.0*PI * sin(0.25 * k)) { + DP K1bm = 0.5 * k - 2.0 * acos (w/(4.0*PI * sin(0.25 * k))); + DP K1bp = 0.5 * k + 2.0 * acos (w/(4.0*PI * sin(0.25 * k))); + Kdomain.Include (K1bm, K1bp); + } + + if (w < 4.0*PI * cos(0.25 * k)) { + DP K1am = 0.5 * k + PI - 2.0 * acos (w/(4.0*PI * cos(0.25 * k))); + DP K1ap = 0.5 * k + PI + 2.0 * acos (w/(4.0*PI * cos(0.25 * k))); + Kdomain.Include (K1am, K1ap); + } + } + + // Now the exclusions: + if (w < twoPI * sin(0.5*k)) { + DP K2dm = 0.5 * k - acos (w/(twoPI * sin (0.5 * k))); + DP K2dp = 0.5 * k + acos (w/(twoPI * sin (0.5 * k))); + Kdomain.Exclude (K2dm, K2dp); + DP K3cm = K2dm + PI; + DP K3cp = K2dp + PI; + Kdomain.Exclude (K3cm, K3cp); + } + + if (w < twoPI * cos(0.5*k)) { + DP K2cm = 0.5 * k + asin(w/(twoPI * cos(0.5*k))); + DP K2cp = 0.5 * k + PI - asin(w/(twoPI * cos(0.5*k))); + Kdomain.Exclude (K2cm, K2cp); + DP K3em = K2cm + PI; + DP K3ep = K2cp + PI; + Kdomain.Exclude (K3em, K3ep); + } + + //cout << "Kdomain: " << endl << Kdomain << endl; + + // Use (K,W) -> (k-K, w-W) symmetry to restrict to K in [k/2, k/2+PI] + Kdomain.Exclude (0.0, 0.5 * k); + Kdomain.Exclude (0.5 * k + PI, twoPI); + //Kdomain.Exclude (0.5 * k, 0.5 * k + PI); + prefactor *= 2.0; + + //cout << "Kdomain restricted: " << endl << Kdomain << endl; + + DP answer = 0.0; + + for (int idom = 0; idom < Kdomain.Ndomains(); ++idom) + //answer += Integrate_rec_using_table (H_fn, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, max_rec); + //answer += Integrate_rec_using_table (H2_fn, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, max_rec); + //answer += Integrate_rec_using_table (H_fn_alt, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, max_rec); + answer += Integrate_rec_using_table (H_fn_mid, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, max_rec); + + //return(prefactor * Integrate_rec_using_table (H_fn, args_to_H, 2, Itable, 0.0, twoPI, req_prec, max_rec)); + //return(prefactor * Riemann_Integrate_rec_using_table (H_fn, args_to_H, 2, Itable, 0.0, twoPI, 500)); + return (prefactor * answer); + } + + DP SF_4p_kwKW_opt (Vect_DP args, I_table Itable) + { + // Translate: + // args[0] = k; + // args[1] = omega; + // args[2] = req_prec; + // args[3] = Npts_K; + // args[4] = Npts_W; + + DP k = args[0]; + DP omega = args[1]; + DP req_prec = args[2]; + int Npts_K = int(args[3]); + int Npts_W = int(args[4]); + + Vect_DP args_to_H(5); + args_to_H[0] = k; // shift of PI in Bougourzi: because they do FM case. + // We want AFM, so SF_4p (k, omega) is correctly obtained directly from the RHS of their formula. + DP w = 2.0 * omega; // Rescale energies by factor 2 because of definitions of H_XXX (omega: S.S; w: 0.5 * sigma.sigma = 2 S.S) + args_to_H[1] = w; + args_to_H[2] = 0.0; // this is K + args_to_H[3] = JSC::max(1.0e-14, 0.01 * req_prec); + args_to_H[4] = DP(Npts_W); + + if (w > wmax_4p(k) || w < wmin_4p(k)) { + //cout << "w out of bounds in SF_4p: " << "wmin = " << PI * sin(k) << " wmax = " << 4.0 * PI * sin(0.25 * k) << " w = " << w << endl; + return(0.0); + } + + DP prefactor = 2.0 * 0.5 * 4.0 * Compute_C4 (req_prec); // 4 comes from using p1 > p2 & p3 > p4 instead of whole interval. + // 2 from Jacobian |dw/domega| + // 0.5 from S^{zz} = S^{pm}/2 + //DP prefactor = 4.0; + + // Define the K integral domain + Domain Kdomain; + + // First, the inclusions: + if (w <= twoPI * sin(0.5*k)) Kdomain.Include (0.0, twoPI); + + else { + if (w < 4.0*PI * sin(0.25 * k)) { + DP K1bm = 0.5 * k - 2.0 * acos (w/(4.0*PI * sin(0.25 * k))); + DP K1bp = 0.5 * k + 2.0 * acos (w/(4.0*PI * sin(0.25 * k))); + Kdomain.Include (K1bm, K1bp); + } + + if (w < 4.0*PI * cos(0.25 * k)) { + DP K1am = 0.5 * k + PI - 2.0 * acos (w/(4.0*PI * cos(0.25 * k))); + DP K1ap = 0.5 * k + PI + 2.0 * acos (w/(4.0*PI * cos(0.25 * k))); + Kdomain.Include (K1am, K1ap); + } + } + + // Now the exclusions: + if (w < twoPI * sin(0.5*k)) { + DP K2dm = 0.5 * k - acos (w/(twoPI * sin (0.5 * k))); + DP K2dp = 0.5 * k + acos (w/(twoPI * sin (0.5 * k))); + Kdomain.Exclude (K2dm, K2dp); + DP K3cm = K2dm + PI; + DP K3cp = K2dp + PI; + Kdomain.Exclude (K3cm, K3cp); + } + + if (w < twoPI * cos(0.5*k)) { + DP K2cm = 0.5 * k + asin(w/(twoPI * cos(0.5*k))); + DP K2cp = 0.5 * k + PI - asin(w/(twoPI * cos(0.5*k))); + Kdomain.Exclude (K2cm, K2cp); + DP K3em = K2cm + PI; + DP K3ep = K2cp + PI; + Kdomain.Exclude (K3em, K3ep); + } + + //cout << "Kdomain: " << endl << Kdomain << endl; + + // Use (K,W) -> (k-K, w-W) symmetry to restrict to K in [k, k+PI] + Kdomain.Exclude (0.0, 0.5 * k); + Kdomain.Exclude (0.5 * k + PI, twoPI); + //Kdomain.Exclude (0.5 * k, 0.5 * k + PI); + prefactor *= 2.0; + + //cout << "Kdomain restricted: " << endl << Kdomain << endl; + + DP answer = 0.0; + + for (int idom = 0; idom < Kdomain.Ndomains(); ++idom) + //answer += Integrate_rec_using_table (H_fn, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, max_rec); + //answer += Integrate_rec_using_table (H2_fn, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, max_rec); + //answer += Integrate_rec_using_table (H_fn_alt, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, max_rec); + //answer += Integrate_rec_using_table (H_fn_mid, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, max_rec); + //answer += Integrate_rec_using_table (H_fn_mid, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, 1); + answer += (Integrate_optimal_using_table (H_fn_mid_opt, args_to_H, 2, Itable, Kdomain.xmin(idom), Kdomain.xmax(idom), req_prec, 1.0e-32, Npts_K)).integ_est; + + //return(prefactor * Integrate_rec_using_table (H_fn, args_to_H, 2, Itable, 0.0, twoPI, req_prec, max_rec)); + //return(prefactor * Riemann_Integrate_rec_using_table (H_fn, args_to_H, 2, Itable, 0.0, twoPI, 500)); + return (prefactor * answer); + } + + DP SF_4p_kwKW_alpha (Vect_DP args, I_table Itable) + { + // Better version for fixed k sum rule + + // Translate: + // args[0] = k; + // args[1] = alpha; <-- integration variable, omega = omegamin + (omegamax - omegamin) (1-cos(alpha)) + // args[2] = req_prec; + // args[3] = max_rec; + + Vect_DP args_to_SF_4p_kwKW = args; + DP omegamin = 0.5 * wmin_4p (args[0]); + DP omegamax = 0.5 * wmax_4p (args[0]); + + args_to_SF_4p_kwKW[1] = omegamin + (omegamax - omegamin) * (1.0 - cos(args[1])); + + return((omegamax - omegamin) * sin(args[1]) * SF_4p_kwKW (args_to_SF_4p_kwKW, Itable)); + } + + DP SF_4p_kwKW_alpha_opt (Vect_DP args, I_table Itable) + { + // Better version for fixed k sum rule + + // Translate: + // args[0] = k; + // args[1] = alpha; <-- integration variable, omega = omegamin + (omegamax - omegamin) (1-cos(alpha)) + // args[2] = req_prec; + // args[3] = Npts_K; + // args[4] = Npts_W; + + Vect_DP args_to_SF_4p_kwKW = args; + DP omegamin = 0.5 * wmin_4p (args[0]); + DP omegamax = 0.5 * wmax_4p (args[0]); + + args_to_SF_4p_kwKW[1] = omegamin + (omegamax - omegamin) * (1.0 - cos(args[1])); + + return((omegamax - omegamin) * sin(args[1]) * SF_4p_kwKW_opt (args_to_SF_4p_kwKW, Itable)); + } + + DP SF_4p_kwKW_cosh_alpha_opt (Vect_DP args, I_table Itable) + { + // Better version for fixed k sum rule + + // Translate: + // args[0] = k; + // args[1] = alpha; <-- integration variable, omega = omegamin * cosh(alpha) + // args[2] = req_prec; + // args[3] = Npts_K; + // args[4] = Npts_W; + + Vect_DP args_to_SF_4p_kwKW = args; + DP omegamin = 0.5 * wmin_4p (args[0]); + //DP omegamax = 0.5 * wmax_4p (args[0]); + + args_to_SF_4p_kwKW[1] = omegamin * cosh(args[1]); + + //return((omegamax - omegamin) * sin(args[1]) * SF_4p_kwKW_opt (args_to_SF_4p_kwKW, Itable)); + return(omegamin * sinh(args[1]) * SF_4p_kwKW_opt (args_to_SF_4p_kwKW, Itable)); + } + + /******************************************************************************************/ + + // Interface to used version: + DP SF_4p_rec (DP k, DP omega, DP req_prec, int max_rec, I_table Itable) + { + // CAREFUL !! This is S(k, omega) = 2 S(k, w) + + Vect_DP args_to_SF(4); + args_to_SF[0] = k; + args_to_SF[1] = omega; + args_to_SF[2] = req_prec; + args_to_SF[3] = DP(max_rec); + + return(2.0 * SF_4p_kwKW (args_to_SF, Itable)); + } + + DP SF_4p (DP k, DP omega, I_table Itable) + { + // Fixes req_prec and max_rec to default values + return(SF_4p_rec (k, omega, default_req_prec, default_max_rec, Itable)); + } + + // Interface to used version: + DP SF_4p_opt (DP k, DP omega, DP req_prec, int Npts_K, int Npts_W, I_table Itable) + { + // CAREFUL !! This is S(k, omega) = 2 S(k, w) + + Vect_DP args_to_SF(5); + args_to_SF[0] = k; + args_to_SF[1] = omega; + args_to_SF[2] = req_prec; + args_to_SF[3] = DP(Npts_K); + args_to_SF[4] = DP(Npts_W); + + return(2.0 * SF_4p_kwKW_opt (args_to_SF, Itable)); + } + + /******************************************************************************************/ + + void Translate_raw_4p_data (DP k, int dim_w, const char* SFraw_Cstr, const char* SF_Cstr, const char* SFsrc_Cstr, I_table Itable) + { + DP omegamin = 0.5 * wmin_4p (k); // Correct for factor of 2 in E between me & Bougourzi + DP omegamax = 0.5 * wmax_4p (k); + + DP alpha_in; + DP SF_in; + DP alpha_in_old = -1.0; + DP SF_in_old = -1.0; + + //int dim_w = int(pow(3.0, max_rec_w + 2)); + DP* alpha = new DP[dim_w]; + DP* omega = new DP[dim_w]; + DP* SF_4p_dat = new DP[dim_w]; + DP* SF_2p_dat = new DP[dim_w]; + int* index = new int[dim_w]; + + ifstream SFraw; + SFraw.open(SFraw_Cstr); + if (SFraw.fail()) JSCerror("Couldn't open SFraw file in Translate."); + + int i = 0; + for (i = 0; i < dim_w; ++i) { + + if (SFraw.peek() == EOF) JSCerror("Not enough data points in file..."); + + index[i] = i; + + SFraw >> alpha_in >> SF_in; + + alpha[i] = alpha_in; + omega[i] = omegamin + (omegamax - omegamin) * (1.0 - cos(alpha_in)); + + // CAREFUL !!! SF_in is S (k, w), and we want S (k, omega) = 2 S(k, w) + SF_4p_dat[i] = 2.0 * SF_in/((omegamax - omegamin) * sin(alpha_in)); + + SF_2p_dat[i] = SF_2p (k, omega[i], Itable); // This already is S(k, omega) + + alpha_in_old = alpha_in; + SF_in_old = SF_in; + } + + SFraw.close(); + + if (i != dim_w) { + JSCerror("Incorrect number of data points in file."); + } + + QuickSort (omega, index, 0, dim_w - 1); + + //for (int j = 0; j < dim_w; ++j) cout << j << "\t" << omega[j] << "\t" << index[j] << endl; + + DP fixed_k_sr_2p = 0.0; + DP fixed_k_sr_4p = 0.0; + DP full_sr_2p = 0.0; + DP full_sr_4p = 0.0; + DP Jac_dalpha = 0.0; // This is domega = (omegamax - omegamin) sin alpha dalpha + + ofstream SF; + SF.open(SF_Cstr); + + for (int j = 0; j < dim_w; ++j) + SF << setprecision(16) << omega[j] << "\t" << SF_4p_dat[index[j] ] << "\t" << SF_2p_dat[index[j] ] << "\t" + << SF_4p_dat[index[j] ] + SF_2p_dat[index[j] ] << endl; + + SF.close(); + + // Compute first moment sum rule + Jac_dalpha = (omegamax - omegamin) * sin(alpha[index[1] ]) * 0.5 * (alpha[index[2] ] - alpha[index[0] ]); + fixed_k_sr_4p += Jac_dalpha * (omega[0] * SF_4p_dat[index[0] ] + omega[1] * SF_4p_dat[index[1] ]); + fixed_k_sr_2p += Jac_dalpha * (omega[0] * SF_2p_dat[index[0] ] + omega[1] * SF_2p_dat[index[1] ]); + full_sr_4p += Jac_dalpha * (SF_4p_dat[index[0] ] + SF_4p_dat[index[1] ]); + full_sr_2p += Jac_dalpha * (SF_2p_dat[index[0] ] + SF_2p_dat[index[1] ]); + + for (int i = 2; i < dim_w - 2; ++i) { + Jac_dalpha = (omegamax - omegamin) * sin(alpha[index[i] ]) * 0.5 * (alpha[index[i + 1] ] - alpha[index[i - 1] ]); + fixed_k_sr_4p += Jac_dalpha * omega[i] * SF_4p_dat[index[i] ]; + fixed_k_sr_2p += Jac_dalpha * omega[i] * SF_2p_dat[index[i] ]; + full_sr_4p += Jac_dalpha * SF_4p_dat[index[i] ]; + full_sr_2p += Jac_dalpha * SF_2p_dat[index[i] ]; + } + + Jac_dalpha = (omegamax - omegamin) * sin(alpha[index[dim_w - 2] ]) * 0.5 * (alpha[index[dim_w - 1] ] - alpha[index[dim_w - 3] ]); + fixed_k_sr_4p += Jac_dalpha * (omega[dim_w - 2] * SF_4p_dat[index[dim_w - 2] ] + omega[dim_w - 1] * SF_4p_dat[index[dim_w - 1] ]); + fixed_k_sr_2p += Jac_dalpha * (omega[dim_w - 2] * SF_2p_dat[index[dim_w - 2] ] + omega[dim_w - 1] * SF_2p_dat[index[dim_w - 1] ]); + full_sr_4p += Jac_dalpha * (SF_4p_dat[index[dim_w - 2] ] + SF_4p_dat[index[dim_w - 1] ]); + full_sr_2p += Jac_dalpha * (SF_2p_dat[index[dim_w - 2] ] + SF_2p_dat[index[dim_w - 1] ]); + + ofstream SFsrc; + SFsrc.open(SFsrc_Cstr); + + // Reintroduce 1/PI since \int domega/2PI + full_sr_4p /= twoPI; + full_sr_2p /= twoPI; + fixed_k_sr_4p /= twoPI; + fixed_k_sr_2p /= twoPI; + + SFsrc << setprecision(16) << fixed_k_sr_4p << "\t" + << fixed_k_sr_2p << "\t" + << fixed_k_sr_4p + fixed_k_sr_2p << "\t" + << fixed_k_sr_4p/Fixed_k_sumrule_omega(k) << "\t" + << fixed_k_sr_2p/Fixed_k_sumrule_omega(k) << "\t" + << (fixed_k_sr_4p + fixed_k_sr_2p)/Fixed_k_sumrule_omega(k) << endl; + + SFsrc << setprecision(16) << full_sr_4p << "\t" << full_sr_2p << "\t" << full_sr_4p + full_sr_2p << "\t" + << 0.25 * full_sr_4p << "\t" << 0.25 * full_sr_2p << "\t" << 0.25 * (full_sr_4p + full_sr_2p) << endl; + + SFsrc.close(); + + delete[] omega; + delete[] SF_4p_dat; + delete[] SF_2p_dat; + delete[] index; + + return; + } + + void Translate_raw_4p_data_cosh (DP k, int dim_w, const char* SFraw_Cstr, const char* SF_Cstr, const char* SFsrc_Cstr, I_table Itable) + { + // Here, omega = omegamin * cosh(alpha) + + DP omegamin = 0.5 * wmin_4p (k); // Correct for factor of 2 in E between me & Bougourzi + //DP omegamax = 0.5 * wmax_4p (k); + + DP alpha_in; + DP SF_in; + DP alpha_in_old = -1.0; + DP SF_in_old = -1.0; + + //int dim_w = int(pow(3.0, max_rec_w + 2)); + DP* alpha = new DP[dim_w]; + DP* omega = new DP[dim_w]; + DP* SF_4p_dat = new DP[dim_w]; + DP* SF_2p_dat = new DP[dim_w]; + int* index = new int[dim_w]; + + ifstream SFraw; + SFraw.open(SFraw_Cstr); + if (SFraw.fail()) JSCerror("Couldn't open SFraw file in Translate."); + + int i = 0; + for (i = 0; i < dim_w; ++i) { + + if (SFraw.peek() == EOF) JSCerror("Not enough data points in file..."); + + index[i] = i; + + SFraw >> alpha_in >> SF_in; + + alpha[i] = alpha_in; + //omega[i] = omegamin + (omegamax - omegamin) * (1.0 - cos(alpha_in)); + omega[i] = omegamin * cosh(alpha_in); + + // CAREFUL !!! SF_in is S (k, w), and we want S (k, omega) = 2 S(k, w) + //SF_4p_dat[i] = 2.0 * SF_in/((omegamax - omegamin) * sin(alpha_in)); + SF_4p_dat[i] = 2.0 * SF_in/(omegamin * sinh(alpha_in)); + + SF_2p_dat[i] = SF_2p (k, omega[i], Itable); // This already is S(k, omega) + + alpha_in_old = alpha_in; + SF_in_old = SF_in; + } + + SFraw.close(); + + if (i != dim_w) { + JSCerror("Incorrect number of data points in file."); + } + + QuickSort (omega, index, 0, dim_w - 1); + + //for (int j = 0; j < dim_w; ++j) cout << j << "\t" << omega[j] << "\t" << index[j] << endl; + + DP fixed_k_sr_2p = 0.0; + DP fixed_k_sr_4p = 0.0; + DP full_sr_2p = 0.0; + DP full_sr_4p = 0.0; + //DP Jac_dalpha = 0.0; // This is domega = (omegamax - omegamin) sin alpha dalpha + DP Jac_dalpha = 0.0; // This is domega = omegamin sinh alpha dalpha + + ofstream SF; + SF.open(SF_Cstr); + + for (int j = 0; j < dim_w; ++j) + SF << setprecision(16) << omega[j] << "\t" << SF_4p_dat[index[j] ] << "\t" << SF_2p_dat[index[j] ] << "\t" + << SF_4p_dat[index[j] ] + SF_2p_dat[index[j] ] << endl; + + SF.close(); + + // Compute first moment sum rule + //Jac_dalpha = (omegamax - omegamin) * sin(alpha[index[1] ]) * 0.5 * (alpha[index[2] ] - alpha[index[0] ]); + Jac_dalpha = omegamin * sinh(alpha[index[1] ]) * 0.5 * (alpha[index[2] ] - alpha[index[0] ]); + fixed_k_sr_4p += Jac_dalpha * (omega[0] * SF_4p_dat[index[0] ] + omega[1] * SF_4p_dat[index[1] ]); + fixed_k_sr_2p += Jac_dalpha * (omega[0] * SF_2p_dat[index[0] ] + omega[1] * SF_2p_dat[index[1] ]); + full_sr_4p += Jac_dalpha * (SF_4p_dat[index[0] ] + SF_4p_dat[index[1] ]); + full_sr_2p += Jac_dalpha * (SF_2p_dat[index[0] ] + SF_2p_dat[index[1] ]); + + for (int i = 2; i < dim_w - 2; ++i) { + //Jac_dalpha = (omegamax - omegamin) * sin(alpha[index[i] ]) * 0.5 * (alpha[index[i + 1] ] - alpha[index[i - 1] ]); + Jac_dalpha = omegamin * sinh(alpha[index[i] ]) * 0.5 * (alpha[index[i + 1] ] - alpha[index[i - 1] ]); + fixed_k_sr_4p += Jac_dalpha * omega[i] * SF_4p_dat[index[i] ]; + fixed_k_sr_2p += Jac_dalpha * omega[i] * SF_2p_dat[index[i] ]; + full_sr_4p += Jac_dalpha * SF_4p_dat[index[i] ]; + full_sr_2p += Jac_dalpha * SF_2p_dat[index[i] ]; + } + + //Jac_dalpha = (omegamax - omegamin) * sin(alpha[index[dim_w - 2] ]) * 0.5 * (alpha[index[dim_w - 1] ] - alpha[index[dim_w - 3] ]); + Jac_dalpha = omegamin * sinh(alpha[index[dim_w - 2] ]) * 0.5 * (alpha[index[dim_w - 1] ] - alpha[index[dim_w - 3] ]); + fixed_k_sr_4p += Jac_dalpha * (omega[dim_w - 2] * SF_4p_dat[index[dim_w - 2] ] + omega[dim_w - 1] * SF_4p_dat[index[dim_w - 1] ]); + fixed_k_sr_2p += Jac_dalpha * (omega[dim_w - 2] * SF_2p_dat[index[dim_w - 2] ] + omega[dim_w - 1] * SF_2p_dat[index[dim_w - 1] ]); + full_sr_4p += Jac_dalpha * (SF_4p_dat[index[dim_w - 2] ] + SF_4p_dat[index[dim_w - 1] ]); + full_sr_2p += Jac_dalpha * (SF_2p_dat[index[dim_w - 2] ] + SF_2p_dat[index[dim_w - 1] ]); + + ofstream SFsrc; + SFsrc.open(SFsrc_Cstr); + + // Reintroduce 1/PI since \int domega/2PI + full_sr_4p /= twoPI; + full_sr_2p /= twoPI; + fixed_k_sr_4p /= twoPI; + fixed_k_sr_2p /= twoPI; + + SFsrc << setprecision(16) << fixed_k_sr_4p << "\t" + << fixed_k_sr_2p << "\t" + << fixed_k_sr_4p + fixed_k_sr_2p << "\t" + << fixed_k_sr_4p/Fixed_k_sumrule_omega(k) << "\t" + << fixed_k_sr_2p/Fixed_k_sumrule_omega(k) << "\t" + << (fixed_k_sr_4p + fixed_k_sr_2p)/Fixed_k_sumrule_omega(k) << endl; + + SFsrc << setprecision(16) << full_sr_4p << "\t" << full_sr_2p << "\t" << full_sr_4p + full_sr_2p << "\t" + << 0.25 * full_sr_4p << "\t" << 0.25 * full_sr_2p << "\t" << 0.25 * (full_sr_4p + full_sr_2p) << endl; + + SFsrc.close(); + + delete[] omega; + delete[] SF_4p_dat; + delete[] SF_2p_dat; + delete[] index; + + return; + } + + + // Function producing a fixed k scan, with data file + DP SF_4p_rec (DP k, DP req_prec, int max_rec_w, int max_rec, I_table Itable) + { + stringstream SFraw_stringstream; + string SFraw_string; + SFraw_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_max_rec_w_" << max_rec_w << "_max_rec_" << max_rec << ".raw"; + SFraw_string = SFraw_stringstream.str(); + const char* SFraw_Cstr = SFraw_string.c_str(); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_max_rec_w_" << max_rec_w << "_max_rec_" << max_rec << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + stringstream SFsrc_stringstream; + string SFsrc_string; + SFsrc_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_max_rec_w_" << max_rec_w << "_max_rec_" << max_rec << ".src"; + SFsrc_string = SFsrc_stringstream.str(); + const char* SFsrc_Cstr = SFsrc_string.c_str(); + + ofstream SFraw_outfile; + SFraw_outfile.open(SFraw_Cstr); + ofstream SFsrc_outfile; + SFsrc_outfile.open(SFsrc_Cstr, ofstream::app); + + //DP omegamin_used = 0.5 * wmin_4p (k); // Correct for factor of 2 in E between me & Bougourzi + //DP omegamax_used = 0.5 * wmax_4p (k); + + Vect_DP args_to_SF(4); + args_to_SF[0] = k; + args_to_SF[1] = 0.0; // integration variable + args_to_SF[2] = req_prec; + args_to_SF[3] = DP(max_rec); + + //DP answer = Integrate_rec_using_table (SF_4p_kwKW, args_to_SF, 1, Itable, omegamin_used, omegamax_used, req_prec, max_rec, SF_outfile); + // Version using omega = omegamin + (omegamax - omegamin) * (1-cos(alpha)) + DP answer = Integrate_rec_using_table (SF_4p_kwKW_alpha, args_to_SF, 1, Itable, 0.0, 0.5*PI, req_prec, max_rec_w, SFraw_outfile)/twoPI; + + SFraw_outfile.close(); + + SFsrc_outfile << answer << endl; + SFsrc_outfile.close(); + + // Translate raw data into SF_4p (k,omega) data + + Translate_raw_4p_data (k, int(pow(3.0, max_rec_w + 2)), SFraw_Cstr, SF_Cstr, SFsrc_Cstr, Itable); + + return(answer); + } + + Integral_result SF_4p_opt (DP k, DP req_prec, int Npts_w, int Npts_K, int Npts_W, I_table Itable) + { + stringstream SFraw_stringstream; + string SFraw_string; + SFraw_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << ".raw"; + //SFraw_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << "_ch.raw"; + SFraw_string = SFraw_stringstream.str(); + const char* SFraw_Cstr = SFraw_string.c_str(); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << ".dat"; + //SF_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << "_ch.dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + stringstream SFsrc_stringstream; + string SFsrc_string; + SFsrc_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << ".src"; + //SFsrc_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << "_ch.src"; + SFsrc_string = SFsrc_stringstream.str(); + const char* SFsrc_Cstr = SFsrc_string.c_str(); + + ofstream SFraw_outfile; + SFraw_outfile.open(SFraw_Cstr); + ofstream SFsrc_outfile; + SFsrc_outfile.open(SFsrc_Cstr); + + //DP omegamin_used = 0.5 * wmin_4p (k); // Correct for factor of 2 in E between me & Bougourzi + //DP omegamax_used = 0.5 * wmax_4p (k); + + Vect_DP args_to_SF(5); + args_to_SF[0] = k; + args_to_SF[1] = 0.0; // integration variable + args_to_SF[2] = req_prec; + args_to_SF[3] = DP(Npts_K); + args_to_SF[4] = DP(Npts_W); + + // Version using omega = omegamin + (omegamax - omegamin) * (1-cos(alpha)) + Integral_result answer = Integrate_optimal_using_table (SF_4p_kwKW_alpha_opt, args_to_SF, 1, Itable, 0.0, 0.5*PI, req_prec, 1.0e-32, Npts_w, SFraw_outfile); + // Version using omega = omegamin * cosh(alpha) + //Integral_result answer = Integrate_optimal_using_table (SF_4p_kwKW_cosh_alpha_opt, args_to_SF, 1, Itable, 0.0, acosh(wmax_4p(k)/wmin_4p(k)), req_prec, 1.0e-32, Npts_w, SFraw_outfile); + answer.integ_est /= twoPI; + answer.abs_prec /= twoPI; + + SFraw_outfile.close(); + + SFsrc_outfile << answer << endl; + SFsrc_outfile.close(); + + // Translate raw data into SF_4p (k,omega) data + + Translate_raw_4p_data (k, answer.n_vals, SFraw_Cstr, SF_Cstr, SFsrc_Cstr, Itable); + //Translate_raw_4p_data_cosh (k, answer.n_vals, SFraw_Cstr, SF_Cstr, SFsrc_Cstr, Itable); + + return(answer); + } + + Integral_result SF_4p_opt (DP k, DP req_prec, int Npts_w, int Npts_KW, I_table Itable) + { + return(SF_4p_opt (k, req_prec, Npts_w, Npts_KW, Npts_KW, Itable)); + } + + Integral_result SF_4p_opt (DP k, DP req_prec, int Npts, I_table Itable) + { + return(SF_4p_opt (k, req_prec, Npts, Npts, Npts, Itable)); + } + + //******************************** Functions to produce files similar to ABACUS ********************************** + + void Produce_SF_2p_file (int N, int Nomega, DP omegamax, I_table Itable) + { + // IMPORTANT NOTE: this produces a file with Szz !! + + DP ABACUS_factor = 1.0; + + if (N % 2) JSCerror("Please use N even in Produce_SF_2p_file."); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_2p_N_" << N << "_Nw_" << Nomega << "_wmax_" << omegamax << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + Write_K_File (N, 0, N); + Write_Omega_File (Nomega, 0.0, omegamax); + + int dim_K = N/2 + 1; + DP* K = new DP[dim_K]; + for (int iK = 0; iK < dim_K; ++iK) K[iK] = (twoPI * iK)/N; + + DP* omega = new DP[Nomega]; + for (int iw = 0; iw < Nomega; ++iw) omega[iw] = omegamax * (iw + 0.5)/Nomega; + + DP* SF_2p_dat = new DP[dim_K * Nomega]; + + DP srtot = 0.0; + Vect_DP sr1(0.0, dim_K); + + for (int iK = 0; iK < dim_K; ++iK) + for (int iw = 0; iw < Nomega; ++iw) { + SF_2p_dat[dim_K * iw + iK] = ABACUS_factor * SF_2p (K[iK], omega[iw], Itable); + srtot += (iK == N/2 ? 1.0 : 2.0) * SF_2p_dat[dim_K * iw + iK]; + sr1[iK] += omega[iw] * SF_2p_dat[dim_K * iw + iK]; + } + + ofstream SF_outfile; + SF_outfile.open(SF_Cstr); + SF_outfile.precision(14); + + for (int iw = 0; iw < Nomega; ++iw) { + for (int iK = 0; iK < dim_K; ++iK) + SF_outfile << SF_2p_dat[dim_K * iw + iK] << "\t"; + for (int iKt = dim_K - 2; iKt >= 0; --iKt) // put K in [PI, 2PI] back in + SF_outfile << SF_2p_dat[dim_K * iw + iKt] << "\t"; + SF_outfile << endl; + } + + SF_outfile.close(); + + // Do sum rule files: + + stringstream SRC_stringstream; + string SRC_string; + SRC_stringstream << "SF_2p_N_" << N << "_Nw_" << Nomega << "_wmax_" << omegamax << ".src"; + SRC_string = SRC_stringstream.str(); + const char* SRC_Cstr = SRC_string.c_str(); + + ofstream SRC_outfile; + SRC_outfile.open(SRC_Cstr); + SRC_outfile.precision(14); + + SRC_outfile << srtot * omegamax/(twoPI * Nomega * N) << "\t" << srtot * 4.0 * omegamax/(twoPI * Nomega * N) << endl; + + SRC_outfile.close(); + + stringstream SR1_stringstream; + string SR1_string; + SR1_stringstream << "SF_2p_N_" << N << "_Nw_" << Nomega << "_wmax_" << omegamax << ".sr1"; + SR1_string = SR1_stringstream.str(); + const char* SR1_Cstr = SR1_string.c_str(); + + ofstream SR1_outfile; + SR1_outfile.open(SR1_Cstr); + SR1_outfile.precision(14); + + for (int iK = 1; iK < dim_K; ++iK) + SR1_outfile << iK << "\t" << K[iK] << "\t" << sr1[iK] * omegamax/(twoPI * Nomega) + << "\t" << -((1.0 - cos(K[iK])) * 2.0 * (0.25 - log(2.0))/3.0) << "\t" + << -sr1[iK] * omegamax/(twoPI * Nomega)/((1.0 - cos(K[iK])) * 2.0 * (0.25 - log(2.0))/3.0) << endl; + + SR1_outfile.close(); + + return; + } + + void Produce_SF_4p_file (int N, int Nomega, DP omegamax, DP req_prec, int max_rec, I_table Itable) + { + // IMPORTANT NOTE: this produces a file with the same normalization as Smp from ABACUS, + // so we use a factor of 2 (for Szz -> Smp) and 2PI. + + DP ABACUS_factor = 1.0; + + if (N % 2) JSCerror("Please use N even in Produce_SF_2p_file."); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_4p_N_" << N << "_Nw_" << Nomega << "_wmax_" << omegamax << "_prec_" << req_prec << "_max_rec_" << max_rec << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + Write_K_File (N, 0, N); + Write_Omega_File (Nomega, 0.0, omegamax); + + int dim_K = N/2 + 1; + DP* K = new DP[dim_K]; + for (int iK = 0; iK < dim_K; ++iK) K[iK] = (twoPI * iK)/N; + + DP* omega = new DP[Nomega]; + for (int iw = 0; iw < Nomega; ++iw) omega[iw] = omegamax * (iw + 0.5)/Nomega; + + DP* SF_4p_dat = new DP[dim_K * Nomega]; + + DP srtot = 0.0; + Vect_DP sr1(0.0, dim_K); + + for (int iK = 0; iK < dim_K; ++iK) + for (int iw = 0; iw < Nomega; ++iw) { + SF_4p_dat[dim_K * iw + iK] = ABACUS_factor * SF_4p_rec (K[iK], omega[iw], req_prec, max_rec, Itable); + srtot += (iK == N/2 ? 1.0 : 2.0) * SF_4p_dat[dim_K * iw + iK]; + sr1[iK] += omega[iw] * SF_4p_dat[dim_K * iw + iK]; + } + + ofstream SF_outfile; + SF_outfile.open(SF_Cstr); + SF_outfile.precision(14); + + for (int iw = 0; iw < Nomega; ++iw) { + for (int iK = 0; iK < dim_K; ++iK) + SF_outfile << SF_4p_dat[dim_K * iw + iK] << "\t"; + for (int iKt = dim_K - 2; iKt >= 0; --iKt) // put K in [PI, 2PI] back in + SF_outfile << SF_4p_dat[dim_K * iw + iKt] << "\t"; + SF_outfile << endl; + } + + SF_outfile.close(); + + // Do sum rule files: + + stringstream SRC_stringstream; + string SRC_string; + SRC_stringstream << "SF_4p_N_" << N << "_Nw_" << Nomega << "_wmax_" << omegamax << "_prec_" << req_prec + << "_max_rec_" << max_rec << ".src"; + SRC_string = SRC_stringstream.str(); + const char* SRC_Cstr = SRC_string.c_str(); + + ofstream SRC_outfile; + SRC_outfile.open(SRC_Cstr); + SRC_outfile.precision(14); + + SRC_outfile << srtot * omegamax/(twoPI * Nomega * N) << "\t" << srtot * 4.0 * omegamax/(twoPI * Nomega * N) << endl; + + SRC_outfile.close(); + + stringstream SR1_stringstream; + string SR1_string; + SR1_stringstream << "SF_4p_N_" << N << "_Nw_" << Nomega << "_wmax_" << omegamax << "_prec_" << req_prec + << "_max_rec_" << max_rec << ".sr1"; + SR1_string = SR1_stringstream.str(); + const char* SR1_Cstr = SR1_string.c_str(); + + ofstream SR1_outfile; + SR1_outfile.open(SR1_Cstr); + SR1_outfile.precision(14); + + for (int iK = 1; iK < dim_K; ++iK) + SR1_outfile << iK << "\t" << K[iK] << "\t" << sr1[iK] * omegamax/(twoPI * Nomega) + << "\t" << -((1.0 - cos(K[iK])) * 2.0 * (0.25 - log(2.0))/3.0) << "\t" + << -sr1[iK] * omegamax/(twoPI * Nomega)/((1.0 - cos(K[iK])) * 2.0 * (0.25 - log(2.0))/3.0) << endl; + + SR1_outfile.close(); + + return; + } + + void Produce_SF_4p_file (int N, int Nomega, DP omegamax, DP req_prec, int Npts_K, int Npts_W, I_table Itable) + { + // IMPORTANT NOTE: this produces a file with the same normalization as Smp from ABACUS, + // so we use a factor of 2 (for Szz -> Smp) and 2PI. + + DP ABACUS_factor = 1.0; + + if (N % 2) JSCerror("Please use N even in Produce_SF_2p_file."); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_4p_N_" << N << "_Nw_" << Nomega << "_wmax_" << omegamax << "_prec_" << req_prec + << "_Npts_K_" << Npts_K << "_Npts_W_" << Npts_W << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + Write_K_File (N, 0, N); + Write_Omega_File (Nomega, 0.0, omegamax); + + int dim_K = N/2 + 1; + DP* K = new DP[dim_K]; + for (int iK = 0; iK < dim_K; ++iK) K[iK] = (twoPI * iK)/N; + + DP* omega = new DP[Nomega]; + for (int iw = 0; iw < Nomega; ++iw) omega[iw] = omegamax * (iw + 0.5)/Nomega; + + DP* SF_4p_dat = new DP[dim_K * Nomega]; + + DP srtot = 0.0; + Vect_DP sr1(0.0, dim_K); + + for (int iK = 0; iK < dim_K; ++iK) + for (int iw = 0; iw < Nomega; ++iw) { + SF_4p_dat[dim_K * iw + iK] = ABACUS_factor * SF_4p_opt (K[iK], omega[iw], req_prec, Npts_K, Npts_W, Itable); + srtot += (iK == N/2 ? 1.0 : 2.0) * SF_4p_dat[dim_K * iw + iK]; + sr1[iK] += omega[iw] * SF_4p_dat[dim_K * iw + iK]; + } + + // Output SF: + ofstream SF_outfile; + SF_outfile.open(SF_Cstr); + SF_outfile.precision(14); + + for (int iw = 0; iw < Nomega; ++iw) { + for (int iK = 0; iK < dim_K; ++iK) + SF_outfile << SF_4p_dat[dim_K * iw + iK] << "\t"; + for (int iKt = dim_K - 2; iKt >= 0; --iKt) // put K in [PI, 2PI] back in + SF_outfile << SF_4p_dat[dim_K * iw + iKt] << "\t"; + SF_outfile << endl; + } + + SF_outfile.close(); + + // Do sum rule files: + + stringstream SRC_stringstream; + string SRC_string; + SRC_stringstream << "SF_4p_N_" << N << "_Nw_" << Nomega << "_wmax_" << omegamax << "_prec_" << req_prec + << "_Npts_K_" << Npts_K << "_Npts_W_" << Npts_W << ".src"; + SRC_string = SRC_stringstream.str(); + const char* SRC_Cstr = SRC_string.c_str(); + + ofstream SRC_outfile; + SRC_outfile.open(SRC_Cstr); + SRC_outfile.precision(14); + + SRC_outfile << srtot * omegamax/(twoPI * Nomega * N) << "\t" << srtot * 4.0 * omegamax/(twoPI * Nomega * N) << endl; + + SRC_outfile.close(); + + stringstream SR1_stringstream; + string SR1_string; + SR1_stringstream << "SF_4p_N_" << N << "_Nw_" << Nomega << "_wmax_" << omegamax << "_prec_" << req_prec + << "_Npts_K_" << Npts_K << "_Npts_W_" << Npts_W << ".sr1"; + SR1_string = SR1_stringstream.str(); + const char* SR1_Cstr = SR1_string.c_str(); + + ofstream SR1_outfile; + SR1_outfile.open(SR1_Cstr); + SR1_outfile.precision(14); + + for (int iK = 1; iK < dim_K; ++iK) + SR1_outfile << iK << "\t" << K[iK] << "\t" << sr1[iK] * omegamax/(twoPI * Nomega) + << "\t" << -((1.0 - cos(K[iK])) * 2.0 * (0.25 - log(2.0))/3.0) << "\t" + << -sr1[iK] * omegamax/(twoPI * Nomega)/((1.0 - cos(K[iK])) * 2.0 * (0.25 - log(2.0))/3.0) << endl; + + SR1_outfile.close(); + + return; + } + + + + + +} // namespace JSC + + diff --git a/src/XXX_h0/XXX_h0_par.cc b/src/XXX_h0/XXX_h0_par.cc new file mode 100755 index 0000000..4751020 --- /dev/null +++ b/src/XXX_h0/XXX_h0_par.cc @@ -0,0 +1,85 @@ +/**************************************************************** + +This software is part of J.-S. Caux's C++ library. + +Copyright (c) 2006. + +----------------------------------------------------------- + +XXX_h0_par.cc + +Zero field structure factor for XXX, 4 spinons, parallel (MPI). + +LAST MODIFIED: 01/11/06 + +******************************************************************/ + +#include "mpi.h" +#include "JSC.h" + +using namespace std; + +namespace JSC { + + // ***************** THIS IS NOT FINISHED !!!!!!!!!!!!!!!!!!!! + + + Integral_result SF_4p_opt_par (MPI_Comm comm, DP k, DP req_prec, int Npts_w, int Npts_K, int Npts_W, I_table Itable) + { + stringstream SFraw_stringstream; + string SFraw_string; + SFraw_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << ".raw"; + //SFraw_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << "_ch.raw"; + SFraw_string = SFraw_stringstream.str(); + const char* SFraw_Cstr = SFraw_string.c_str(); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << ".dat"; + //SF_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << "_ch.dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + stringstream SFsrc_stringstream; + string SFsrc_string; + SFsrc_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << ".src"; + //SFsrc_stringstream << "SF_4p_k_" << k << "_prec_" << req_prec << "_Npts_" << Npts_w << "_" << Npts_K << "_" << Npts_W << "_ch.src"; + SFsrc_string = SFsrc_stringstream.str(); + const char* SFsrc_Cstr = SFsrc_string.c_str(); + + ofstream SFraw_outfile; + SFraw_outfile.open(SFraw_Cstr); + ofstream SFsrc_outfile; + SFsrc_outfile.open(SFsrc_Cstr); + + //DP omegamin_used = 0.5 * wmin_4p (k); // Correct for factor of 2 in E between me & Bougourzi + //DP omegamax_used = 0.5 * wmax_4p (k); + + Vect_DP args_to_SF(5); + args_to_SF[0] = k; + args_to_SF[1] = 0.0; // integration variable + args_to_SF[2] = req_prec; + args_to_SF[3] = DP(Npts_K); + args_to_SF[4] = DP(Npts_W); + + // Version using omega = omegamin + (omegamax - omegamin) * (1-cos(alpha)) + Integral_result answer = Integrate_optimal_using_table (SF_4p_kwKW_alpha_opt, args_to_SF, 1, Itable, 0.0, 0.5*PI, req_prec, 1.0e-32, Npts_w, SFraw_outfile); + // Version using omega = omegamin * cosh(alpha) + //Integral_result answer = Integrate_optimal_using_table (SF_4p_kwKW_cosh_alpha_opt, args_to_SF, 1, Itable, 0.0, acosh(wmax_4p(k)/wmin_4p(k)), req_prec, 1.0e-32, Npts_w, SFraw_outfile); + answer.integ_est /= twoPI; + answer.abs_prec /= twoPI; + + SFraw_outfile.close(); + + SFsrc_outfile << answer << endl; + SFsrc_outfile.close(); + + // Translate raw data into SF_4p (k,omega) data + + Translate_raw_4p_data (k, answer.n_vals, SFraw_Cstr, SF_Cstr, SFsrc_Cstr, Itable); + //Translate_raw_4p_data_cosh (k, answer.n_vals, SFraw_Cstr, SF_Cstr, SFsrc_Cstr, Itable); + + return(answer); + } + +} // namespace JSC diff --git a/src/XXX_h0/XXX_h0_v2.cc b/src/XXX_h0/XXX_h0_v2.cc new file mode 100755 index 0000000..dd1cc5a --- /dev/null +++ b/src/XXX_h0/XXX_h0_v2.cc @@ -0,0 +1,232 @@ +/**************************************************************** + +XXX_h0_v2.cc + +Defines all class procedures used for the XXX chain in zero field. +Aims at calculating the full k, omega dependence in one go. + +LAST MODIFIED: 19/02/07 + +******************************************************************/ + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + + DP Direct_J_integral (int Npts_p, DP req_prec, I_table Itable) + { + stringstream outfile_stringstream; + string outfile_string; + outfile_stringstream << "SF_4p_Np_" << Npts_p << "_prec_" << req_prec << ".raw"; + outfile_string = outfile_stringstream.str(); + const char* outfile_Cstr = outfile_string.c_str(); + + ofstream outfile; + outfile.open(outfile_Cstr); + outfile.precision(16); + + stringstream src_stringstream; + string src_string; + src_stringstream << "SF_4p_Np_" << Npts_p << "_prec_" << req_prec << ".src"; + src_string = src_stringstream.str(); + const char* src_Cstr = src_string.c_str(); + + ofstream srcfile; + srcfile.open(src_Cstr); + srcfile.precision(16); + + Vect_DP p(4); + DP J_fn_cont = 0.0; + DP sum_J_fn = 0.0; + for (int ip0 = 0; ip0 < Npts_p; ++ip0) { + p[0] = - (ip0 + 0.5) * PI/Npts_p; + for (int ip1 = ip0 + 1; ip1 < Npts_p; ++ip1) { + p[1] = - (ip1 + 0.5) * PI/Npts_p; + for (int ip2 = ip1 + 1; ip2 < Npts_p; ++ip2) { + p[2] = - (ip2 + 0.5) * PI/Npts_p; + for (int ip3 = ip2 + 1; ip3 < Npts_p; ++ip3) { + p[3] = - (ip3 + 0.5) * PI/Npts_p; + J_fn_cont = J_fn (p, req_prec, Itable); + sum_J_fn += J_fn_cont; + if (p[0] > 0 || p[1] > 1 || p[2] > 2 || p[3] > 3) outfile << endl; + outfile << p << "\t" << J_fn_cont; + } + } + } + } + + DP integral_J = Compute_C4 (req_prec) * 24.0 * sum_J_fn * pow(PI/Npts_p, 4.0); + // 24: from using ordered p's + + srcfile << integral_J << "\t" << integral_J/(4.0 * PI * PI) << "\t" << integral_J/(PI * PI) << endl; + + outfile.close(); + srcfile.close(); + + return(integral_J); + } + + void Smoothen_raw_SF_4p (int Npts_p, int Npts_o, DP req_prec, DP width) + { + stringstream rawfile_stringstream; + string rawfile_string; + rawfile_stringstream << "SF_4p_Np_" << Npts_p << "_No_" << Npts_o << "_prec_" << req_prec << ".raw"; + rawfile_string = rawfile_stringstream.str(); + const char* rawfile_Cstr = rawfile_string.c_str(); + + ifstream rawfile; + rawfile.open(rawfile_Cstr); + + stringstream smfile_stringstream; + string smfile_string; + smfile_stringstream << "SF_4p_Np_" << Npts_p << "_No_" << Npts_o << "_prec_" << req_prec << "_w_" << width << ".dat"; + smfile_string = smfile_stringstream.str(); + const char* smfile_Cstr = smfile_string.c_str(); + + ofstream smfile; + smfile.open(smfile_Cstr); + smfile.precision(16); + + DP* SF_4p_raw = new DP[(Npts_p + 1) * Npts_o]; + DP* SF_4p_sm = new DP[(Npts_p + 1) * Npts_o]; + + for (int iw = 0; iw < Npts_o; ++iw) + for (int ik = 0; ik <= Npts_p; ++ik) + rawfile >> SF_4p_raw[Npts_o * ik + iw]; + + rawfile.close(); + + // We now broaden to a Gaussian: + + int iw_half_window = int(width/(2.0 * PI/Npts_o)); + if (iw_half_window == 0) JSCerror("width too small in Smoothen_raw_SF_4p"); + + int iwstart, iwstop; + DP domega_sq_over_width_sq = pow(2.0 * PI/(Npts_o * width), 2.0); + DP exp_factor; + + for (int iw = 0; iw < Npts_o; ++iw) { + iwstart = JSC::max(0, iw - 10 * iw_half_window); + iwstop = JSC::min(Npts_o, iw + 10 * iw_half_window); + + for (int iwraw = iwstart; iwraw <= iwstop; ++iwraw) { + exp_factor = exp(-(iwraw - iw) * (iwraw - iw) * domega_sq_over_width_sq); + for (int ik = 0; ik <= Npts_p; ++ik) + { + SF_4p_sm[Npts_o * ik + iw] += SF_4p_raw[Npts_o * ik + iwraw] * exp_factor; + } + } + } + + // Reset normalization from delta function and output data + + DP prefactor = 1.0/(sqrt(PI) * width); + for (int iw = 0; iw < Npts_o; ++iw) { + for (int ik = 0; ik <= Npts_p; ++ik) + smfile << SF_4p_sm[Npts_o * ik + iw] * prefactor << "\t"; + smfile << endl; + } + + smfile.close(); + } + + DP Direct_J_integral_bin (int Npts_p, int Npts_o, DP req_prec, I_table Itable) + { + // Produces binned file for J + + stringstream rawfile_stringstream; + string rawfile_string; + rawfile_stringstream << "SF_4p_Np_" << Npts_p << "_No_" << Npts_o << "_prec_" << req_prec << ".raw"; + rawfile_string = rawfile_stringstream.str(); + const char* rawfile_Cstr = rawfile_string.c_str(); + + ofstream rawfile; + rawfile.open(rawfile_Cstr); + rawfile.precision(16); + + stringstream src_stringstream; + string src_string; + src_stringstream << "SF_4p_Np_" << Npts_p << "_No_" << Npts_o << "_prec_" << req_prec << ".src"; + src_string = src_stringstream.str(); + const char* src_Cstr = src_string.c_str(); + + ofstream srcfile; + srcfile.open(src_Cstr); + srcfile.precision(16); + + Vect_DP p(4); + DP J_fn_cont = 0.0; + DP sum_J_fn = 0.0; + + DP* SF_4p_raw = new DP[(Npts_p + 1) * Npts_o]; + // labelling: index is Npts_o * ik + iomega + + DP omegamax = 2.0 * PI; + + Heis_Write_K_File (Npts_p); + Heis_Write_w_File (Npts_o, omegamax); + + int ik; + DP omega; + int iomega; + DP sinp0, sinp1, sinp2; + + for (int ip0 = 0; ip0 < Npts_p; ++ip0) { + p[0] = - (ip0 + 0.5) * PI/Npts_p; + sinp0 = sin(p[0]); + for (int ip1 = ip0 + 1; ip1 < Npts_p; ++ip1) { + p[1] = - (ip1 + 0.5) * PI/Npts_p; + sinp1 = sin(p[1]); + for (int ip2 = ip1 + 1; ip2 < Npts_p; ++ip2) { + p[2] = - (ip2 + 0.5) * PI/Npts_p; + sinp2 = sin(p[2]); + for (int ip3 = ip2 + 1; ip3 < Npts_p; ++ip3) { + p[3] = - (ip3 + 0.5) * PI/Npts_p; + + ik = ip0 + ip1 + ip2 + ip3; + if (ik > 2 * Npts_p) ik -= 2*Npts_p; // guarantees 0 <= ik < 2Npts + + if (ik <= Npts_p) { + + omega = -0.5 * PI * (sinp0 + sinp1 + sinp2 + sin(p[3])); + iomega = int(omega * Npts_o/omegamax); + + //cout << ik << "\t" << iomega << endl; + + J_fn_cont = J_fn (p, req_prec, Itable); + sum_J_fn += (ik == Npts_p ? 1.0 : 2.0) * J_fn_cont; + + SF_4p_raw[Npts_o * ik + iomega] += J_fn_cont; + } + } + } + } + } + + DP prefactor = Compute_C4 (req_prec) * 24.0 * pow(PI/Npts_p, 4.0); + // 24: from using ordered p's + + DP integral_J = prefactor * sum_J_fn; + + // Output raw file + for (int iw = 0; iw < Npts_o; ++iw) { + for (int iK = 0; iK <= Npts_p; ++iK) rawfile << prefactor * SF_4p_raw[Npts_o * iK + iw] << "\t"; + rawfile << endl; + } + + srcfile << integral_J << "\t" << integral_J/(4.0 * PI * PI) << "\t" << integral_J/(PI * PI) << endl; + + rawfile.close(); + srcfile.close(); + + delete[] SF_4p_raw; + + return(integral_J); + } + +} // namespace JSC + + diff --git a/src/XXZ_h0/XXZ_h0.cc b/src/XXZ_h0/XXZ_h0.cc new file mode 100644 index 0000000..3d692b3 --- /dev/null +++ b/src/XXZ_h0/XXZ_h0.cc @@ -0,0 +1,701 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) + +----------------------------------------------------------- + +File: XXZ_h0.cc + +Purpose: Defines all class procedures used for the infinite XXZ chain in zero field. + + +***********************************************************/ + + +#include "JSC.h" + +using namespace std; + +namespace JSC { + + DP Fermi_velocity_XXZ_h0 (DP Delta) + { + // This uses the J > 0 conventions, with H = J \sum S.S + + if (Delta <= -1.0 || Delta >= 1.0) JSCerror("Use Delta in AFM gapless regime for Fermi_velocity_XXZ_h0."); + + return(0.5 * PI * sqrt(1.0 - Delta * Delta)/acos(Delta)); + } + + inline DP Integrand_xi_11 (Vect_DP args) + { + // args[0] corresponds to t, args[1] to rho, args[2] to xi + + + // For increased accuracy: separate in regions 0 < t <= 1 and 1 < t, and 0 < xi t <= 1 and 1 < xi t: + + DP answer; + + if (args[0] * args[2] <= 1.0) { + + if (args[0] <= 1.0) { + answer = sinh(args[0] * (1.0 + args[2])) * sinh(args[0]) * cos(4.0 * args[0] * args[1]) + /(args[0] * sinh(args[0] * args[2]) * pow(cosh(args[0]), 2.0)); + } + else { // factorize exp t + DP expm2t = exp(-2.0*args[0]); + answer = sinh(args[0] * (1.0 + args[2])) * 2.0 * (1.0 - expm2t) * cos(4.0 * args[0] * args[1]) + /(args[0] * sinh(args[0] * args[2]) * (1.0 + expm2t) * (1.0 + expm2t)); + } + } + + else { // factorize exp (xi t) + if (args[0] <= 1.0) { + DP expm2t = exp(-2.0 * args[0]); + DP expm2txi = exp(-2.0*args[0]*args[2]); + answer = (1.0 - expm2t * expm2txi) * sinh(args[0]) * cos(4.0 * args[0] * args[1]) + /(args[0] * (1.0 - expm2txi) * 0.5 * (1.0 + expm2t) * cosh(args[0])); + } + else { // factorize exp t + DP expm2t = exp(-2.0 * args[0]); + DP expm2txi = exp(-2.0*args[0]*args[2]); + answer = 2.0 * (1.0 - expm2t * expm2txi) * (1.0 - expm2t) * cos(4.0 * args[0] * args[1]) + /(args[0] * (1.0 - expm2txi) * (1.0 + expm2t) * (1.0 + expm2t)); + } + } + + return(answer); + + //return(sinh(args[0]*(1.0 + args[2])) * sinh(args[0]) * cos(4.0 * args[0] * args[1])/(args[0] * sinh(args[0] * args[2]) * pow(cosh(args[0]), 2.0))); + } + + DP I_xi_11_integral (DP xi, DP rho, DP req_prec, int max_nr_pts, DP t1bar) + { + DP rho_used = fabs(rho); + + Vect_DP args(3); + args[0] = 0.0; + args[1] = rho_used; + args[2] = xi; + + return((Integrate_optimal (Integrand_xi_11, args, 0, 0.0, t1bar, req_prec, req_prec, max_nr_pts)).integ_est); + } + + inline DP Integrand_xi_12 (Vect_DP args) + { + // args[0] corresponds to t, args[1] to rho, args[2] to xi + DP expm2t = exp(-2.0*args[0]); + DP expm2txi = exp(-2.0*args[0]*args[2]); + + return(2.0 * cos(4.0 * args[0] * args[1]) * (-expm2t * (3.0 + expm2t) + expm2txi * (1.0 + expm2t * (1.0 + 2.0*expm2t))) + / (args[0] * (1.0 - expm2txi) * (1.0 + expm2t) * (1.0 + expm2t))); + } + + DP I_xi_12_integral (DP xi, DP rho, DP req_prec, int max_nr_pts, DP t1bar) + { + DP tinf = 24.0; // such that exp(-2tinf) << machine_eps + DP rho_used = fabs(rho); + + Vect_DP args(3); + args[0] = 0.0; + args[1] = rho_used; + args[2] = xi; + + return((Integrate_optimal (Integrand_xi_12, args, 0, t1bar, tinf, req_prec, req_prec, max_nr_pts)).integ_est); + } + + /* + inline DP Integrand_xi_2 (Vect_DP args) + { + // This version is used for rho <= 1 + + // args[0] corresponds to t, args[1] to rho, args[2] to xi + + DP answer; + + if (args[0] * args[2] <= 1.0) { + + if (args[0] <= 1.0) { + answer = sinh(args[0] * (args[2] + 1.0)) * pow(sin(2.0 * args[0] * args[1]), 2.0) + /(args[0] * sinh(args[2] * args[0]) * sinh(args[0]) * pow(cosh(args[0]), 2.0)); + } + + else if (args[0] >= 1.0) { + DP expm2t = exp(-2.0 * args[0]); + DP expm2txi = exp(-2.0*args[0]*args[2]); + //answer = 8.0 * expm2t * pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * (1.0 - expm2t) * (1.0 + expm2t) * (1.0 + expm2t)); + answer = 8.0 * (1.0 - expm2t*expm2txi) * expm2t * + pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * (1.0 - expm2txi) * (1.0 - expm2t) * (1.0 + expm2t) * (1.0 + expm2t)); + } + + DP expm2t = exp(-2.0 * args[0]); + DP expm2txi = exp(-2.0*args[0]*args[2]); + //answer = 8.0 * expm2t * pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * (1.0 - expm2t) * (1.0 + expm2t) * (1.0 + expm2t)); + answer = 8.0 * (1.0 - expm2t*expm2txi) * expm2t * + pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * (1.0 - expm2txi) * (1.0 - expm2t) * (1.0 + expm2t) * (1.0 + expm2t)); + + return(answer); + + ... NOT USED + + } + */ + + inline DP Integrand_xi_22 (Vect_DP args) + { + // This version is used when rho > 1. + + // args[0] corresponds to t, args[1] to rho, args[2] to xi + DP answer = 0.0; + if (args[0] < 1.0) + answer = pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * tanh(args[0] * args[2]) * pow(cosh(args[0]), 2.0)); + else if (args[0] >= 1.0) { + DP expm2t = exp(-2.0 * args[0]); + DP expm2txi = exp(-2.0*args[0]*args[2]); + answer = 4.0 * expm2t * (1.0 + expm2txi) * pow(sin(2.0 * args[0] * args[1]), 2.0)/(args[0] * (1.0 - expm2txi) * (1.0 + expm2t) * (1.0 + expm2t)); + } + return(answer); + } + + DP I_xi_2_integral (DP xi, DP rho, DP req_prec, int max_nr_pts) + { + DP tinf = 24.0; // such that exp(-2tinf) << machine_eps + DP rho_used = fabs(rho); + + Vect_DP args(3); + args[0] = 0.0; + args[1] = rho_used; + args[2] = xi; + + DP answer = 0.0; + + //if (rho_used <= 1.0) answer = (Integrate_optimal (Integrand_xi_2, args, 0, 0.0, tinf, req_prec, req_prec, max_nr_pts)).integ_est; + + //else + answer = PI * rho + log(0.5 * (1.0 + exp(-2.0 * PI * rho_used))) // This is I^{(21)} + + (Integrate_optimal (Integrand_xi_22, args, 0, 0.0, tinf, req_prec, req_prec, max_nr_pts)).integ_est; + + return(answer); + } + + + DP I_xi_integral (DP xi, DP rho, DP req_prec, int max_nr_pts) + { + + DP t1bar = 1.0; // delimiter between the I_xi_11 and I__xi_12 integrals + + DP answer = I_xi_11_integral (xi, rho, req_prec, max_nr_pts, t1bar) + + I_xi_12_integral (xi, rho, req_prec, max_nr_pts, t1bar) + - 2.0 * Cosine_Integral (4.0 * rho * t1bar) + - I_xi_2_integral (xi, rho, req_prec, max_nr_pts); + + return(answer); + } + + + + /********************* TWO SPINONS ********************/ + + DP Szz_XXZ_h0_2spinons (DP Delta, DP k, DP omega, Integral_table Itable) + { + // Careful ! This is S(k, omega) = S (k, w) |dw/domega| = 2 S(k, w) + + DP w = 2.0 * omega; // Rescale energies by factor 2 because of definitions of H_XXX (omega: S.S; w: 0.5 * sigma.sigma = 2 S.S) + + DP vF = Fermi_velocity_XXZ_h0 (Delta); // in units of omega + + //DP wu = twoPI * sin(0.5 * k); + DP wu = 4.0 * vF * sin(0.5 * k); + //DP wl = PI * fabs(sin(k)); + DP wl = 2.0 * vF * fabs(sin(k)); + + DP rho = acosh(sqrt((wu * wu - wl * wl)/(w * w - wl * wl)))/PI; + DP xi = PI/acos(Delta) - 1.0; + // Factor of 2: return S(k, omega), not S(k, w) + // 0.25 factor: 1/4 * 2 * 1/2, where 1/4 comes from Bougourzi, 2 is the Jacobian |dw/domega| and 1/2 is S^{zz} = 1/2 * S^{+-} + //return(w < wu && w > wl ? 2.0 * 0.5 * exp(-Itable.Return_val (acosh(sqrt((wu * wu - wl * wl)/(w * w - wl * wl)))/PI))/sqrt(wu * wu - w * w) : 0.0); + DP expmtwoPIrhooverxi = exp(-twoPI * rho/xi); + //return(w < wu && w > wl ? 2.0 * exp(-Itable.Return_val (rho))/(sqrt(wu * wu - w * w) * (cosh(twoPI * rho/xi) + cos(PI/xi))) : 0.0); + return(w < wu && w > wl ? 2.0 * pow(1.0 + 1.0/xi, 2.0) * exp(-Itable.Return_val (rho)) + * expmtwoPIrhooverxi/(sqrt(wu * wu - w * w) + * (0.5 * (1.0 + expmtwoPIrhooverxi * expmtwoPIrhooverxi) + expmtwoPIrhooverxi * cos(PI/xi))) + : 0.0); + + } + + DP Szz_XXZ_h0_2spinons (Vect_DP args, Integral_table Itable) + { + // To be called by integrating functions + + // Careful ! This is S(k, omega) ! + + // This uses args[0] = k, args[1] = omega, args[2] = xi + + DP Delta = cos(PI/(args[2] + 1.0)); + + return(Szz_XXZ_h0_2spinons (Delta, args[0], args[1], Itable)); + + /* + + // This uses args[0] = k, args[1] = w, args[2] = xi + + DP Delta = cos(PI/(args[2] + 1.0)); + + DP vF = Fermi_velocity_XXZ_h0 (Delta); // in units of omega + + //DP wu = twoPI * sin(0.5 * k); + DP wu = 4.0 * vF * sin(0.5 * args[0]); + //DP wl = PI * fabs(sin(k)); + DP wl = 2.0 * vF * fabs(sin(args[0])); + + DP rho = acosh(sqrt((wu * wu - wl * wl)/(args[1] * args[1] - wl * wl)))/PI; + DP expmtwoPIrhooverxi = exp(-twoPI * rho/args[2]); + + // 0.5 factor: 1 from Bougourzi, and 1/2 is S^{zz} = 1/2 * S^{+-} + return(args[1] < wu && args[1] > wl ? + //0.5 * exp(-Itable.Return_val (acosh(sqrt((wu * wu - wl * wl)/(args[1] * args[1] - wl * wl)))/PI))/sqrt(wu * wu - args[1] * args[1]) : 0.0); + pow(1.0 + 1.0/args[2], 2.0) * exp(-Itable.Return_val (rho)) * expmtwoPIrhooverxi + /(sqrt(wu * wu - args[1] * args[1]) + * (0.5 * (1.0 + expmtwoPIrhooverxi * expmtwoPIrhooverxi) + expmtwoPIrhooverxi * cos(PI/args[2]))) : 0.0); + + */ + } + + DP Szz_XXZ_h0_2spinons_alt (Vect_DP args, Integral_table Itable) + { + // This returns S(k, omega) \times Jacobian (d omega)/(d alpha) where omega = omega_{lower threshold} cosh \alpha + + // so returns S(k, omega) \sqrt{omega^2 - omega_l^2} + + // This uses args[0] = k, args[1] = alpha, args[2] = xi + + DP Delta = cos(PI/(args[2] + 1.0)); + DP vF = Fermi_velocity_XXZ_h0 (Delta); // in units of omega + + DP omegaup = 2.0 * vF * sin(0.5 * args[0]); + DP omegalow = vF * fabs(sin(args[0])); + + DP omega = omegalow * cosh(args[1]); + + if (omega >= omegaup || omega <= omegalow) return(0.0); + + DP Jacobian = sqrt(omega * omega - omegalow * omegalow); + + return(Jacobian * Szz_XXZ_h0_2spinons (Delta, args[0], omega, Itable)); + + /* + //DP wu = twoPI * sin(0.5 * k); + DP wu = 4.0 * vF * sin(0.5 * args[0]); + //DP wl = PI * fabs(sin(k)); + DP wl = 2.0 * vF * fabs(sin(args[0])); + + //DP w = wu * cos(args[1]); + //DP factor = 1.0; + DP w = wl * cosh(args[1]); + + if (w >= wu || w <= wl) return(0.0); + + DP factor = sqrt((w * w - wl * wl)/(wu * wu - w * w)); + + DP rho = acosh(sqrt((wu * wu - wl * wl)/(w * w - wl * wl)))/PI; + DP expmtwoPIrhooverxi = exp(-twoPI * rho/args[2]); + + // 0.5 factor: 1 from Bougourzi, and 1/2 is S^{zz} = 1/2 * S^{+-} + //return(factor * 0.5 * exp(-Itable.Return_val (acosh(sqrt((wu * wu - wl * wl)/(w * w - wl * wl)))/PI))); + return(pow(1.0 + 1.0/args[2], 2.0) * factor + * exp(-Itable.Return_val (rho)) * expmtwoPIrhooverxi/(0.5 * (1.0 + expmtwoPIrhooverxi * expmtwoPIrhooverxi) + expmtwoPIrhooverxi * cos(PI/args[2]))); + */ + + } + + DP Szz_XXZ_h0_2spinons_omega (Vect_DP args, Integral_table Itable) + { + // This uses args[0] = k, args[1] = omega, args[2] = xi + + return(args[1] * Szz_XXZ_h0_2spinons (args, Itable)); + } + + DP Szz_XXZ_h0_2spinons_omega_alt (Vect_DP args, Integral_table Itable) + { + // This uses args[0] = k, args[1] = alpha, args[2] = xi + + DP Delta = cos(PI/(args[2] + 1.0)); + DP vF = Fermi_velocity_XXZ_h0 (Delta); // in units of omega + + DP omegalow = vF * fabs(sin(args[0])); + + DP omega = omegalow * cosh(args[1]); + + args[1] = omega; + + return(Szz_XXZ_h0_2spinons_omega (args, Itable)); + } + + + DP Szz_XXZ_h0_2spinons_intomega (Vect_DP args, Integral_table Itable) + { + // This returns \int_0^2PI domega/2PI S(k, omega) + + DP k = args[0]; + DP req_prec = args[1]; + DP xi = args[2]; + int max_nr_pts = int(args[3]); + + DP Delta = cos(PI/(args[2] + 1.0)); + DP vF = Fermi_velocity_XXZ_h0 (Delta); // in units of omega + + DP omegaup = 2.0 * vF * sin(0.5 * args[0]); + DP omegalow = vF * fabs(sin(args[0])); + + Vect_DP args_to_SF_2p(4); + args_to_SF_2p[0] = k; + args_to_SF_2p[1] = 0.0; // this will be omega + args_to_SF_2p[2] = xi; + args_to_SF_2p[3] = JSC::max(1.0e-14, 0.01 * req_prec); + + return((Integrate_optimal_using_table (Szz_XXZ_h0_2spinons, args_to_SF_2p, 1, Itable, + omegalow, omegaup, req_prec, req_prec, max_nr_pts)).integ_est/twoPI); + } + + DP Szz_XXZ_h0_2spinons_intomega_alt (Vect_DP args, Integral_table Itable) + { + // This returns \int_0^2PI domega/2PI S(k, omega) + + DP k = args[0]; + DP req_prec = args[1]; + DP xi = args[2]; + int max_nr_pts = int(args[3]); + + DP Delta = cos(PI/(args[2] + 1.0)); + DP vF = Fermi_velocity_XXZ_h0 (Delta); // in units of omega + + DP omegaup = 2.0 * vF * sin(0.5 * args[0]); + DP omegalow = vF * fabs(sin(args[0])); + + Vect_DP args_to_SF_2p(4); + args_to_SF_2p[0] = k; + args_to_SF_2p[1] = 0.0; // this will be omega + args_to_SF_2p[2] = xi; + args_to_SF_2p[3] = JSC::max(1.0e-14, 0.01 * req_prec); + + //return(Integrate_rec_using_table (SF_2p_alt, args_to_SF_2p, 1, Itable, 0.0, acos(wl/wu), req_prec, max_rec)/twoPI); + return((Integrate_optimal_using_table (Szz_XXZ_h0_2spinons_alt, args_to_SF_2p, 1, Itable, + 0.0, acosh(omegaup/omegalow), req_prec, req_prec, max_nr_pts)).integ_est/twoPI); + } + + DP Szz_XXZ_h0_2spinons_check_sumrule (DP Delta, DP req_prec, int max_nr_pts, Integral_table Itable) + { + // Near XX: it's better to use this function. + // Near XXX: it's better to use the ..._alt function below. + + Vect_DP args_to_Szz_intomega (4); + + args_to_Szz_intomega[0] = 0.0; // this will be k + args_to_Szz_intomega[1] = JSC::max(1.0e-14, 0.01 * req_prec); + args_to_Szz_intomega[2] = PI/acos(Delta) - 1.0; + args_to_Szz_intomega[3] = DP(max_nr_pts); + + // Factor 4: we expect \int_0^\infty \frac{d\omega}{2\pi} \int_0^{2\pi} \frac{dk}{2\pi} S^zz (k, omega) = 1/4 + // Factor 2: int[0, 2PI] = 2 int[0, PI] + return(4.0 * 2.0 * (Integrate_optimal_using_table (Szz_XXZ_h0_2spinons_intomega, args_to_Szz_intomega, 0, Itable, + 0.0, PI, req_prec, req_prec, max_nr_pts)).integ_est/twoPI); + } + + DP Szz_XXZ_h0_2spinons_check_sumrule_alt (DP Delta, DP req_prec, int max_nr_pts, Integral_table Itable) + { + // This is the preferred version if near XXX. + + Vect_DP args_to_Szz_intomega (4); + + args_to_Szz_intomega[0] = 0.0; // this will be k + args_to_Szz_intomega[1] = JSC::max(1.0e-14, 0.01 * req_prec); + args_to_Szz_intomega[2] = PI/acos(Delta) - 1.0; + args_to_Szz_intomega[3] = DP(max_nr_pts); + + // Factor 2: int[0, 2PI] = 2 int[0, PI] + return(4.0 * 2.0 * (Integrate_optimal_using_table (Szz_XXZ_h0_2spinons_intomega_alt, args_to_Szz_intomega, 0, Itable, + 0.0, PI, req_prec, req_prec, max_nr_pts)).integ_est/twoPI); + } + + DP Fixed_k_sumrule_omega_Szz_XXZ_h0_N (DP Delta, DP k) + { + // This is K_1 (k) = \int domega/2PI omega S(k, omega) + //return(4.0 * (log(2.0) - 0.25) * (1.0 - cos(k))/3.0); // XXX_h0 + + // Calculate the average: + int N = 600; + int M = 300; + DP Xavg = X_avg ('x', Delta, N, M); + + return (-2.0 * (1.0 - cos(k)) * Xavg/N); + } + + DP Integrand_GSE_XXZ_h0 (Vect_DP args) + { + // args[0] corresponds to t, args[1] to xi + DP expm2t = exp(-2.0 * args[0]); + DP expm2txi = exp(-2.0 * args[0] * args[1]); + + return(2.0 * expm2t * (1.0 - expm2txi)/((1.0 - expm2t * expm2txi) * (1.0 + expm2t))); + } + + DP GSE_XXZ_h0 (DP Delta, DP req_prec, int max_nr_pts) + { + DP xi = PI/acos(Delta) - 1.0; + + DP tinf = 24.0; // such that exp(-2tinf) << machine_eps + + Vect_DP args(2); + args[0] = 0.0; + args[1] = xi; + + return(-0.25 * sqrt(1.0 - Delta * Delta) * (2.0/acos(Delta)) * 2.0 * (Integrate_optimal (Integrand_GSE_XXZ_h0, args, 0, 0.0, tinf, req_prec, req_prec, max_nr_pts).integ_est)); + } + + DP Integrand_2_fSR_XXZ_h0 (Vect_DP args) + { + // args[0] corresponds to t, args[1] to xi + DP expm2t = exp(-2.0 * args[0]); + DP expm2txi = exp(-2.0 * args[0] * args[1]); + + return(4.0 * args[0] * expm2t * (1.0 + expm2t * expm2txi) + /((1.0 - expm2t * expm2txi) * (1.0 + expm2t) * (1.0 + expm2t))); + } + + DP Fixed_k_sumrule_omega_Szz_XXZ_h0 (DP Delta, DP k, DP req_prec, int max_nr_pts) + { + // + DP xi = PI/acos(Delta) - 1.0; + + DP tinf = 24.0; // such that exp(-2tinf) << machine_eps + + Vect_DP args(2); + args[0] = 0.0; + args[1] = xi; + + DP Xavg = -(0.25/(acos(Delta) * sqrt(1.0 - Delta * Delta))) * 2.0 * (Integrate_optimal (Integrand_GSE_XXZ_h0, args, 0, 0.0, tinf, req_prec, req_prec, max_nr_pts).integ_est) + + 0.25 * (Delta/pow(acos(Delta), 2.0)) * 2.0 * (Integrate_optimal (Integrand_2_fSR_XXZ_h0, args, 0, 0.0, tinf, req_prec, req_prec, max_nr_pts).integ_est); + + return (-2.0 * (1.0 - cos(k)) * Xavg); + } + + DP Szz_XXZ_h0_2spinons_check_fixed_k_Szz_sumrule (DP Delta, DP k, DP req_prec, int max_nr_pts, Integral_table Itable) + { + // Near XXX, it's better to use the ..._alt function below. + + DP vF = Fermi_velocity_XXZ_h0 (Delta); // in units of omega + + DP omegaup = 2.0 * vF * sin(0.5 * k); + DP omegalow = vF * fabs(sin(k)); + + Vect_DP args_to_SF_2p(4); + args_to_SF_2p[0] = k; + args_to_SF_2p[1] = 0.0; // this will be omega + args_to_SF_2p[2] = PI/acos(Delta) - 1.0; + args_to_SF_2p[3] = JSC::max(1.0e-14, 0.01 * req_prec); + + return(((Integrate_optimal_using_table (Szz_XXZ_h0_2spinons_omega, args_to_SF_2p, 1, Itable, omegalow, omegaup, req_prec, req_prec, max_nr_pts)).integ_est/twoPI)/Fixed_k_sumrule_omega_Szz_XXZ_h0(Delta, k, req_prec, max_nr_pts)); + } + + DP Szz_XXZ_h0_2spinons_check_fixed_k_Szz_sumrule_alt (DP Delta, DP k, DP req_prec, int max_nr_pts, Integral_table Itable) + { + // This is the preferred version if near XXX. + + DP vF = Fermi_velocity_XXZ_h0 (Delta); // in units of omega + + DP omegaup = 2.0 * vF * sin(0.5 * k); + DP omegalow = vF * fabs(sin(k)); + + Vect_DP args_to_SF_2p(4); + args_to_SF_2p[0] = k; + args_to_SF_2p[1] = 0.0; // this will be alpha + args_to_SF_2p[2] = PI/acos(Delta) - 1.0; + args_to_SF_2p[3] = JSC::max(1.0e-14, 0.01 * req_prec); + + return(((Integrate_optimal_using_table (Szz_XXZ_h0_2spinons_omega_alt, args_to_SF_2p, 1, Itable, 0.0, acosh(omegaup/omegalow), req_prec, req_prec, max_nr_pts)).integ_est/twoPI)/Fixed_k_sumrule_omega_Szz_XXZ_h0(Delta, k, req_prec, max_nr_pts)); + } + + + //******************************** Functions to produce files similar to ABACUS ********************************** + + void Produce_Szz_XXZ_h0_2spinons_file (DP Delta, int N, int Nomega, DP omegamax, Integral_table Itable) + { + // IMPORTANT NOTE: this produces a file with Szz !! + + if (N % 2) JSCerror("Please use N even in Produce_Szz_XXZ_h0_2spinons_file."); + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "Szz_XXZ_h0_2spinons_Delta_" << Delta << "_N_" << N << "_Nom_" << Nomega << "_ommax_" << omegamax << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + Write_K_File (N, 0, N); + Write_Omega_File (Nomega, 0.0, omegamax); + + int dim_K = N/2 + 1; + DP* K = new DP[dim_K]; + for (int iK = 0; iK < dim_K; ++iK) K[iK] = (twoPI * iK)/N; + + DP* omega = new DP[Nomega]; + for (int iw = 0; iw < Nomega; ++iw) omega[iw] = omegamax * (iw + 0.5)/Nomega; + + DP* SF_XXZ_2p_dat = new DP[dim_K * Nomega]; + + DP srtot = 0.0; + Vect_DP sr1(0.0, dim_K); + + for (int iK = 0; iK < dim_K; ++iK) + for (int iw = 0; iw < Nomega; ++iw) { + SF_XXZ_2p_dat[dim_K * iw + iK] = Szz_XXZ_h0_2spinons (Delta, K[iK], omega[iw], Itable); + srtot += (iK == N/2 ? 1.0 : 2.0) * SF_XXZ_2p_dat[dim_K * iw + iK]; + sr1[iK] += omega[iw] * SF_XXZ_2p_dat[dim_K * iw + iK]; + } + + ofstream SF_outfile; + SF_outfile.open(SF_Cstr); + SF_outfile.precision(14); + + for (int iw = 0; iw < Nomega; ++iw) { + for (int iK = 0; iK < dim_K; ++iK) + SF_outfile << SF_XXZ_2p_dat[dim_K * iw + iK] << "\t"; + for (int iKt = dim_K - 2; iKt >= 0; --iKt) // put K in [PI, 2PI] back in + SF_outfile << SF_XXZ_2p_dat[dim_K * iw + iKt] << "\t"; + SF_outfile << endl; + } + + SF_outfile.close(); + + // Do sum rule files: + + stringstream SRC_stringstream; + string SRC_string; + SRC_stringstream << "Szz_XXZ_h0_2spinons_Delta_" << Delta << "_N_" << N << "_Nom_" << Nomega << "_ommax_" << omegamax << ".src"; + SRC_string = SRC_stringstream.str(); + const char* SRC_Cstr = SRC_string.c_str(); + + ofstream SRC_outfile; + SRC_outfile.open(SRC_Cstr); + SRC_outfile.precision(14); + + SRC_outfile << srtot * omegamax/(twoPI * Nomega * N) << "\t" << srtot * 4.0 * omegamax/(twoPI * Nomega * N) << endl; + + SRC_outfile.close(); + + stringstream SR1_stringstream; + string SR1_string; + SR1_stringstream << "Szz_XXZ_h0_2spinons_Delta_" << Delta << "_N_" << N << "_Nom_" << Nomega << "_ommax_" << omegamax << ".sr1"; + SR1_string = SR1_stringstream.str(); + const char* SR1_Cstr = SR1_string.c_str(); + + ofstream SR1_outfile; + SR1_outfile.open(SR1_Cstr); + SR1_outfile.precision(14); + + // Figure out the f-sumrule factor: + int Nfsr = 600; + int Mfsr = 300; + DP Xavg = 0.0; + if (Delta > 0.0) Xavg = X_avg ('x', Delta, Nfsr, Mfsr); + else Xavg = 0.0; + + DP sr1factor = 1.0; + if (Delta > 0.0) sr1factor = -2.0 * Xavg/Nfsr; + else sr1factor = 1.0; + + for (int iK = 1; iK < dim_K; ++iK) + SR1_outfile << iK << "\t" << K[iK] << "\t" << sr1[iK] * omegamax/(twoPI * Nomega) + //<< "\t" << -((1.0 - cos(K[iK])) * 2.0 * (0.25 - log(2.0))/3.0) << "\t" + << "\t" << (1.0 - cos(K[iK])) * sr1factor << "\t" + //<< -sr1[iK] * omegamax/(twoPI * Nomega)/((1.0 - cos(K[iK])) * 2.0 * (0.25 - log(2.0))/3.0) << endl; + << sr1[iK] * omegamax/(twoPI * Nomega)/((1.0 - cos(K[iK])) * sr1factor) << endl; + + SR1_outfile.close(); + + return; + } + + void Produce_Szz_XXZ_h0_2spinons_fixed_K_file (DP Delta, DP Kover2PI, int Nomega, Integral_table Itable) + { + // IMPORTANT NOTE: this produces a file with Szz !! + + stringstream SF_stringstream; + string SF_string; + SF_stringstream << "Szz_XXZ_h0_2spinons_Delta_" << Delta << "_Kover2PI_" << Kover2PI + << "_Nom_" << Nomega << ".dat"; + SF_string = SF_stringstream.str(); + const char* SF_Cstr = SF_string.c_str(); + + DP K = twoPI * Kover2PI; + + DP* omega = new DP[Nomega]; + + DP vF = Fermi_velocity_XXZ_h0 (Delta); // in units of omega + + DP omegaup = 2.0 * vF * sin(0.5 * K); + DP omegalow = vF * fabs(sin(K)); + + for (int iw = 0; iw < Nomega; ++iw) omega[iw] = 0.99 * omegalow + (1.01 * omegaup - 0.99 * omegalow) * (iw + 0.5)/Nomega; // factor of 1.01: just to have zeroes on either side of continuum. + + DP* SF_XXZ_2p_dat = new DP[Nomega]; + + DP sr1 = 0.0; + + for (int iw = 0; iw < Nomega; ++iw) { + SF_XXZ_2p_dat[iw] = Szz_XXZ_h0_2spinons (Delta, K, omega[iw], Itable); + sr1 += omega[iw] * SF_XXZ_2p_dat[iw]; + } + + ofstream SF_outfile; + SF_outfile.open(SF_Cstr); + SF_outfile.precision(14); + + SF_outfile << omega[0] << "\t" << SF_XXZ_2p_dat[0]; + for (int iw = 0; iw < Nomega; ++iw) { + SF_outfile << endl << omega[iw] << "\t" << SF_XXZ_2p_dat[iw]; + } + + SF_outfile.close(); + + // Do sum rule files: + /* + NOT ACCURATE SINCE WE'RE PLOTTING SOMEWHAT OUTSIDE OF CONTINUUM AS WELL (to get nice figures) + stringstream SR1_stringstream; + string SR1_string; + SR1_stringstream << "Szz_XXZ_h0_2spinons_Delta_" << Delta << "_Kover2PI_" << Kover2PI + << "_Nom_" << Nomega << ".sr1"; + SR1_string = SR1_stringstream.str(); + const char* SR1_Cstr = SR1_string.c_str(); + + ofstream SR1_outfile; + SR1_outfile.open(SR1_Cstr); + SR1_outfile.precision(14); + + // Figure out the f-sumrule factor: + int Nfsr = 600; + int Mfsr = 300; + DP Xavg = 0.0; + if (Delta > 0.0) Xavg = X_avg ('x', Delta, Nfsr, Mfsr); + else Xavg = 0.0; + + DP sr1factor = 1.0; + if (Delta > 0.0) sr1factor = -2.0 * Xavg/Nfsr; + else sr1factor = 1.0; + + SR1_outfile << Kover2PI << "\t" << sr1 * (omegaup - omegalow)/(twoPI * Nomega) + << "\t" << (1.0 - cos(K)) * sr1factor << "\t" + << sr1 * (omegaup - omegalow)/(twoPI * Nomega)/((1.0 - cos(K)) * sr1factor) << endl; + + SR1_outfile.close(); + */ + + return; + } + + + +} // namespace JSC + + diff --git a/src/YOUNG/Young_Tableau.cc b/src/YOUNG/Young_Tableau.cc new file mode 100644 index 0000000..c1c2b62 --- /dev/null +++ b/src/YOUNG/Young_Tableau.cc @@ -0,0 +1,926 @@ +/********************************************************** + +This software is part of J.-S. Caux's ABACUS++ library. + +Copyright (c) 2006-9. + +----------------------------------------------------------- + +File: Young_Tableau.cc + +Purpose: Defines Young_Tableau class and procedures. + +Last modified: 08/10/2009 + +***********************************************************/ + +#include "JSC.h" + +namespace JSC { + + Young_Tableau::Young_Tableau () : Nrows(0), Ncols(0), Row_L(0), Col_L(0), id(0LL), maxid(0LL), + map(new long long int [1]), map_computed(false), idnr_reached(0LL), nboxes_reached(-1), + dimchoose(0), choose_table(0LL) + {} + + Young_Tableau::Young_Tableau (int Nr, int Nc) + : Nrows(Nr), Ncols(Nc), Row_L(new int[Nrows]), Col_L(new int[Ncols]), id(0LL), + maxid(choose_lli(Nr + Nc, Nc) - 1LL), + map(new long long int [YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1]), + map_computed(false), idnr_reached(0LL), nboxes_reached(-1), + dimchoose (JSC::min(Nr, Nc) + 1), + choose_table(new long long int[(Nr + Nc + 1) * dimchoose]) + + { + // Constructs empty tableau of appropriate size + + for (int i = 0; i < Nrows; ++i) Row_L[i] = 0; + for (int i = 0; i < Ncols; ++i) Col_L[i] = 0; + + // Construct the choose_table + + for (int cti = 0; cti < Nr + Nc + 1; ++cti) + for (int ctj = 0; ctj < dimchoose; ++ctj) { + if (cti >= ctj) choose_table[dimchoose * cti + ctj] = choose_lli(cti, ctj); + else choose_table[dimchoose * cti + ctj] = 0LL; + } + + // Fill map with zeros: + for (int i = 0; i < (YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1); ++i) map[i] = 0LL; + + } + /* SEGFAULTS + Young_Tableau::Young_Tableau (int Nr, int Nc, long long int idnr) + : Nrows(Nr), Ncols(Nc), Row_L(new int[Nrows]), Col_L(new int[Ncols]), id(idnr), + maxid(choose_lli(Nr + Nc, Nc) - 1LL), + map(new long long int [YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1]), + map_computed(false), idnr_reached(0LL), nboxes_reached(-1), + dimchoose (JSC::min(Nr, Nc) + 1), + choose_table(new long long int[(Nr + Nc + 1) * dimchoose]) + { + // Constructs Young tableau of given idnr, if consistent with Nr, Nc. + + // Construct the choose_table + for (int cti = 0; cti < Nr + Nc + 1; ++cti) + for (int ctj = 0; ctj < dimchoose; ++ctj) { + if (cti >= ctj) choose_table[dimchoose * cti + ctj] = choose_lli(cti, ctj); + else choose_table[dimchoose * cti + ctj] = 0LL; + } + + for (int i = 0; i < YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1; ++i) map[i] = 0LL; + + (*this).Set_to_id(idnr); + } + */ + Young_Tableau::Young_Tableau (const Young_Tableau& RefTableau) // copy constructor + : Nrows(RefTableau.Nrows), Ncols(RefTableau.Ncols), Row_L(new int[RefTableau.Nrows]), Col_L(new int[RefTableau.Ncols]), + id(RefTableau.id), maxid(RefTableau.maxid), + map(new long long int [YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1]), + map_computed(RefTableau.map_computed), idnr_reached(RefTableau.idnr_reached), + nboxes_reached(RefTableau.nboxes_reached), + dimchoose (RefTableau.dimchoose), + choose_table(new long long int[(Nrows + Ncols + 1) * dimchoose]) + { + for (int i = 0; i < Nrows; ++i) Row_L[i] = RefTableau.Row_L[i]; + for (int i = 0; i < Ncols; ++i) Col_L[i] = RefTableau.Col_L[i]; + + // Construct the choose_table + for (int cti = 0; cti < Nrows + Ncols + 1; ++cti) + for (int ctj = 0; ctj < dimchoose; ++ctj) { + if (cti >= ctj) choose_table[dimchoose * cti + ctj] = choose_lli(cti, ctj); + else choose_table[dimchoose * cti + ctj] = 0LL; + } + + // The map: + for (int i = 0; i < (YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1); ++i) map[i] = RefTableau.map[i]; + + } + /* + Young_Tableau::Young_Tableau (int Nr, int Nc, long long int* ref_choose_table, int dimref) + : Nrows(Nr), Ncols(Nc), Row_L(new int[Nrows]), Col_L(new int[Ncols]), id(0LL), + maxid(choose_lli(Nr + Nc, Nc) - 1LL), + //choose_table(new long long int[(Nr + Nc + 1) * (Nr + Nc + 1)]), + choose_table(new long long int[(Nr + Nc + 1) * (JSC::min(Nr, Nc) + 1)]), + //map(new long long int[JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT)]), + map(new long long int[YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1]), + map_computed(false), idnr_reached(0LL), nboxes_reached(-1) + + { + // Constructs empty tableau of appropriate size + + for (int i = 0; i < Nrows; ++i) Row_L[i] = 0; + for (int i = 0; i < Ncols; ++i) Col_L[i] = 0; + + // Construct the choose_table + + // Copy entries from reference table + for (int cti = 0; cti < JSC::min(Nr + Nc + 1, dimref); ++cti) + for (int ctj = 0; ctj < JSC::min(Nr + Nc + 1, dimref); ++ctj) + choose_table[(Nr + Nc + 1) * cti + ctj] = cti >= ctj ? ref_choose_table[dimref * cti + ctj] : 0LL; + + // add missing parts if there are any + if (dimref < Nr + Nc + 1) { + for (int cti = 0; cti < Nr + Nc + 1; ++cti) + for (int ctj = dimref; ctj < Nr + Nc + 1; ++ctj) + choose_table[(Nr + Nc + 1) * cti + ctj] = 0LL; + for (int cti = dimref; cti < Nr + Nc + 1; ++cti) + for (int ctj = 0; ctj < Nr + Nc + 1; ++ctj) + choose_table[(Nr + Nc + 1) * cti + ctj] = cti >= ctj ? choose_lli(cti, ctj) : 0LL; + } + + // The map: + //for (int i = 0; i < JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT); ++i) map[i] = 0LL; + for (int i = 0; i < (YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1); ++i) map[i] = 0LL; + } + */ + Young_Tableau::Young_Tableau (int Nr, int Nc, const Young_Tableau& RefTableau) + : Nrows(Nr), Ncols(Nc), Row_L(new int[Nrows]), Col_L(new int[Ncols]), id(0LL), + maxid(choose_lli(Nr + Nc, Nc) - 1LL), + map(new long long int[YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1]), + map_computed(false), idnr_reached(0LL), nboxes_reached(-1), + dimchoose (JSC::min(Nr, Nc) + 1), + choose_table(new long long int[(Nr + Nc + 1) * dimchoose]) + { + // Constructs empty tableau of appropriate size + + for (int i = 0; i < Nrows; ++i) Row_L[i] = 0; + for (int i = 0; i < Ncols; ++i) Col_L[i] = 0; + + // Construct the choose_table + + // Copy entries from reference table + for (int cti = 0; cti < JSC::min(Nr + Nc + 1, RefTableau.Nrows + RefTableau.Ncols + 1); ++cti) + for (int ctj = 0; ctj < JSC::min(dimchoose, RefTableau.dimchoose); ++ctj) + choose_table[dimchoose * cti + ctj] = cti >= ctj ? RefTableau.choose_table[RefTableau.dimchoose * cti + ctj] : 0LL; + + // add missing parts if there are any + int refdim1 = RefTableau.Nrows + RefTableau.Ncols + 1; + + for (int cti = 0; cti < Nr + Nc + 1; ++cti) + for (int ctj = 0; ctj < dimchoose; ++ctj) + if (cti >= refdim1 || ctj >= RefTableau.dimchoose) + choose_table[dimchoose * cti + ctj] = cti >= ctj ? choose_lli(cti, ctj) : 0LL; + + // The map: + for (int i = 0; i < (YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1); ++i) map[i] = 0LL; + } + + Young_Tableau& Young_Tableau::operator= (const Young_Tableau& RefTableau) + { + if (this != &RefTableau) { + Nrows = RefTableau.Nrows; + Ncols = RefTableau.Ncols; + if (Row_L != 0) delete[] Row_L; + Row_L = new int[Nrows]; + for (int i = 0; i < Nrows; ++i) Row_L[i] = RefTableau.Row_L[i]; + if (Col_L != 0) delete[] Col_L; + Col_L = new int[Ncols]; + for (int i = 0; i < Ncols; ++i) Col_L[i] = RefTableau.Col_L[i]; + id = RefTableau.id; + maxid = RefTableau.maxid; + dimchoose = RefTableau.dimchoose; + if (choose_table != 0LL) delete[] choose_table; + choose_table = new long long int[(Nrows + Ncols + 1) * dimchoose]; + for (int cti = 0; cti < Nrows + Ncols + 1; ++cti) + for (int ctj = 0; ctj < dimchoose; ++ctj) + { + choose_table[dimchoose * cti + ctj] = RefTableau.choose_table[dimchoose * cti + ctj]; + } + + if (map != 0LL) delete[] map; + map = new long long int[YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1]; + for (long long int i = 0; i < (YOUNG_TABLEAU_ID_OPTION == 2 ? JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT) : 1); ++i) map[i] = RefTableau.map[i]; + map_computed = RefTableau.map_computed; + idnr_reached = RefTableau.idnr_reached; + nboxes_reached = RefTableau.nboxes_reached; + } + + if (choose_table[0] != 1LL) { // compute the table + + for (int cti = 0; cti < Nrows + Ncols + 1; ++cti) + for (int ctj = 0; ctj < dimchoose; ++ctj) + { + choose_table[dimchoose * cti + ctj] = cti >= ctj ? choose_lli(cti, ctj) : 0LL; + } + } + + return(*this); + } + + Young_Tableau::~Young_Tableau () // destructor + { + if (Row_L != 0) delete[] Row_L; + if (Col_L != 0) delete[] Col_L; + if (choose_table != 0) delete[] choose_table; + if (map != 0LL) delete[] map; + } + + //********************************************************************************* + // Member functions + + Young_Tableau& Young_Tableau::Compute_Map (long long int idnr_to_reach) + { + if (idnr_to_reach < 0LL) JSCerror("negative id requested in Compute_Map"); + + long long int idnr_to_reach_here = JSC::min(idnr_to_reach, JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT)); + + if (!map_computed) { + nboxes_reached = -1; + idnr_reached = 0LL; + map_computed = true; + } + + if (idnr_to_reach_here >= idnr_reached) { + + if (maxid == 0) map[0] = 0LL; + + else { + + // Keep a copy of the Tableau, in case... + int* old_Row_L = new int[Nrows]; + for (int i = 0; i < Nrows; ++i) old_Row_L[i] = Row_L[i]; + + while (idnr_reached <= JSC::min(idnr_to_reach, TABLEAU_ID_UPPER_LIMIT)) { + nboxes_reached++; + Distribute_boxes (nboxes_reached, 0); + } + + // Reset Row_L + for (int i = 0; i < Nrows; ++i) Row_L[i] = old_Row_L[i]; + delete[] old_Row_L; + Set_Col_L_given_Row_L(); + } + + } + return(*this); + } + + Young_Tableau& Young_Tableau::Distribute_boxes (int nboxes_to_dist, int level) + { + if (level > Nrows) { + // Failed to achieve proper distribution. Do nothing. + } + + else if (nboxes_to_dist > Ncols * (Nrows - level)) { + cout << Nrows << "\t" << Ncols << "\t" << level << "\t" << nboxes_to_dist << "\t" << idnr_reached << "\t" << nboxes_reached << endl; + JSCerror("nboxes_to_dist too high"); + } + else if (nboxes_to_dist == 0) { + for (int j = level; j < Nrows; ++j) Row_L[j] = 0; + Compute_id (0); + if (idnr_reached < TABLEAU_ID_UPPER_LIMIT) map[idnr_reached] = id; + idnr_reached++; + } + + else { + int maxlength = level == 0 ? Ncols : Row_L[level - 1]; + + for (int nboxes_of_level = (nboxes_to_dist - 1)/(Nrows - level) + 1; + nboxes_of_level <= JSC::min(nboxes_to_dist, maxlength); ++nboxes_of_level) { + Row_L[level] = nboxes_of_level; + Distribute_boxes (nboxes_to_dist - nboxes_of_level, level + 1); + } + } + return(*this); + } + + Young_Tableau& Young_Tableau::Compute_id() + { + Compute_id(0); + return(*this); + } + + long long int Young_Tableau::Compute_Descendent_id (int option, Vect_INT& Desc_Row_L, int Nrows_Desc, int Ncols_Desc, + const Young_Tableau& RefTableau) + { + long long int answer = 0; + + if (option == 0) { + if (Nrows_Desc == 0) answer = 0; + else if (Desc_Row_L[0] == 0) answer = 0; + else if (Nrows_Desc == 1) answer = Desc_Row_L[0]; + else { + int highest_occupied_row = Nrows_Desc - 1; + while (Desc_Row_L[highest_occupied_row] == 0) highest_occupied_row--; // index of highest occupied row; + + for (int j = 0; j < Desc_Row_L[highest_occupied_row]; ++j) + answer += RefTableau.choose_table[RefTableau.dimchoose * (highest_occupied_row + Ncols_Desc - j) + + JSC::min(highest_occupied_row, Ncols_Desc - j)]; + + Vect_INT Desc_Desc_Row_L(highest_occupied_row); + for (int i = 0; i < highest_occupied_row; ++i) Desc_Desc_Row_L[i] = Desc_Row_L[i] - Desc_Row_L[highest_occupied_row]; + + answer += Compute_Descendent_id (0, Desc_Desc_Row_L, highest_occupied_row, Ncols_Desc - Desc_Row_L[highest_occupied_row], + RefTableau); + + } + } + + else if (option == 1) { + + if (Nrows_Desc == 0 || Ncols_Desc == 0) answer = 0; + + else if (Desc_Row_L[0] == 0) answer = 0; + + else if ((Nrows_Desc == 1) || (Desc_Row_L[1] == 0)) answer = Desc_Row_L[0]; + + else { + int ndiag = 0; + while (Desc_Row_L[ndiag] > ndiag) ndiag++; + + if (ndiag == 1) { + int Desc_Col_L_0 = Nrows_Desc; + while (Desc_Row_L[Desc_Col_L_0 - 1] == 0) Desc_Col_L_0--; + answer = (Desc_Col_L_0 - 1) * Ncols_Desc + Desc_Row_L[0]; + } + + else { + + for (int j = 0; j < ndiag; ++j) answer += RefTableau.choose_table[RefTableau.dimchoose * Nrows_Desc + j] + * RefTableau.choose_table[RefTableau.dimchoose * Ncols_Desc + j]; + + Vect_INT Desc1_Row_L(ndiag); + for (int i = 0; i < ndiag; ++i) Desc1_Row_L[i] = Desc_Row_L[i] - ndiag; + + Vect_INT Desc2_Row_L(ndiag); + for (int i = 0; i < ndiag; ++i) { + Desc2_Row_L[i] = 0; + while(Desc_Row_L[Desc2_Row_L[i] ] > i && Desc2_Row_L[i] < Ncols_Desc) Desc2_Row_L[i]++; + Desc2_Row_L[i] -= ndiag; + } + + long long int Desc1_id = Compute_Descendent_id (1, Desc1_Row_L, ndiag, Ncols_Desc - ndiag, RefTableau); + long long int Desc2_id = Compute_Descendent_id (1, Desc2_Row_L, ndiag, Nrows_Desc - ndiag, RefTableau); + answer += Desc2_id * RefTableau.choose_table[RefTableau.dimchoose * Ncols_Desc + ndiag] + Desc1_id; + } + } + + } // else if (option == 1) + + else JSCerror("Wrong option in Young_Tableau::Compute_Descendent_id"); + + return(answer); + } + + Young_Tableau& Young_Tableau::Compute_id (int option) // computes the id according to rule option + { + + long long int idnr = 0; + + if (option == 0) { + + if (Nrows == 0) idnr = 0; + + else if (Row_L[0] == 0) idnr = 0; + + else if (Nrows == 1) idnr = Row_L[0]; + + else { + + int highest_occupied_row = Nrows - 1; + while (Row_L[highest_occupied_row] == 0) highest_occupied_row--; // index of highest occupied row; + + for (int j = 0; j < Row_L[highest_occupied_row]; ++j) idnr += choose_table[dimchoose * (highest_occupied_row + Ncols - j) + + JSC::min(highest_occupied_row, Ncols - j)]; + + Vect_INT Desc_Row_L(highest_occupied_row); + + for (int i = 0; i < highest_occupied_row; ++i) Desc_Row_L[i] = Row_L[i] - Row_L[highest_occupied_row]; + + idnr += Compute_Descendent_id (0, Desc_Row_L, highest_occupied_row, Ncols - Row_L[highest_occupied_row], (*this)); + } + } + + else if (option == 1) { + + if (Nrows == 0 || Ncols == 0) idnr = 0; + + else if (Row_L[0] == 0) idnr = 0; + + else if ((Nrows == 1) || (Row_L[1] == 0)) idnr = Row_L[0]; + + else { + int ndiag = 0; + while (Row_L[ndiag] > ndiag) ndiag++; + + if (ndiag == 1) { + idnr = (Col_L[0] - 1) * Ncols + Row_L[0]; + } + + else { + + for (int j = 0; j < ndiag; ++j) idnr += choose_table[dimchoose * Nrows + j] * choose_table[dimchoose * Ncols + j]; + + Vect_INT Desc1_Row_L(ndiag); + for (int i = 0; i < ndiag; ++i) Desc1_Row_L[i] = Row_L[i] - ndiag; + + Vect_INT Desc2_Row_L(ndiag); + for (int i = 0; i < ndiag; ++i) Desc2_Row_L[i] = Col_L[i] - ndiag; + + long long int Desc1_id = Compute_Descendent_id (1, Desc1_Row_L, ndiag, Ncols - ndiag, (*this)); + long long int Desc2_id = Compute_Descendent_id (1, Desc2_Row_L, ndiag, Nrows - ndiag, (*this)); + idnr += Desc2_id * choose_table[dimchoose * Ncols + ndiag] + Desc1_id; + } + } + + } // else if (option == 1) + + else if (option == 2) { + + // The order here is given by first listing all tableaux with 0 boxes, then 1, 2, ... by using Map + + Compute_id (0); // sets the id according to rule 0 + Compute_Map (idnr); // make sure the state map is computed + while (map[idnr] != id && idnr < JSC::min(maxid + 1LL, TABLEAU_ID_UPPER_LIMIT)) idnr++; // match with inverse map to get the idnr according to rule 2 + } + + else JSCerror("Wrong option for Tableau ids"); + + id = idnr; + + return (*this); + } + + Young_Tableau& Young_Tableau::Set_to_id (long long int idnr) + { + Set_to_id (idnr, 0); + return(*this); + } + + Young_Tableau& Young_Tableau::Set_to_id (long long int idnr, int option) // sets the tableau to the one corresponding to idnr + { + + if (option == 0) { + + if ((idnr < 0) || ((maxid < idnr) && (Nrows*Ncols != 0))) { + cout << "Nrows = " << Nrows << "\tNcols = " << Ncols << "\tmaxid = " << maxid << "\trequested id = " << idnr << endl; + JSCerror("Wrong idnr in Set_to_id for Young Tableau."); + } + id = idnr; + + int NColumnseff = Ncols; + long long int ideff = idnr; + + for (int i = 0; i < Ncols; ++i) { + + if (ideff == 0) Col_L[i] = 0; + + else { + + int nbar = 0; + + if (ideff <= 0) nbar = 0; + + while (ideff >= choose_table[dimchoose * (NColumnseff + nbar) + JSC::min(NColumnseff, nbar)]) nbar++; + + Col_L[i] = nbar; + + ideff -= choose_table[dimchoose * (NColumnseff + Col_L[i] - 1) + JSC::min(NColumnseff, Col_L[i] - 1)]; + NColumnseff--; + + } + + } // for (int i = 0; i < Ncols... + + (*this).Set_Row_L_given_Col_L(); + + } // if (option == 0) + + else if (option == 1) { + + if ((idnr < 0LL) || ((maxid < idnr) && (Nrows*Ncols != 0))) JSCerror("Wrong idnr in Set_to_id for Young Tableau."); + + if (Nrows*Ncols == 0 && idnr != 0LL) JSCerror("Trying nonzero id on empty Tableau."); + + id = idnr; + + if (idnr == 0LL) { + for (int i = 0; i < Nrows; ++i) Row_L[i] = 0; + for (int i = 0; i < Ncols; ++i) Col_L[i] = 0; + } + + else { + int ndiag = 0; + int sum = 0; + while ((sum < idnr) && (ndiag < Nrows) && (ndiag < Ncols)) { + ndiag++; + sum += choose_table[dimchoose * Nrows + ndiag] * choose_table[dimchoose * Ncols + ndiag]; + } + + long long int residual_id = idnr - 1 - sum + choose_table[dimchoose * Nrows + ndiag] * choose_table[dimchoose * Ncols + ndiag]; + + if (ndiag == 0 && idnr != 0LL) JSCerror("Zero ndiag for nonzero idnr in Tableau."); + + else if (ndiag == 1) { + if (Nrows >= Ncols) { + Col_L[0] = (idnr - 1)/Ncols + 1; + for (int i = 1; i < idnr - Ncols * (Col_L[0] - 1); ++i) Col_L[i] = 1; + for (int i = idnr - Ncols * (Col_L[0] - 1); i < Ncols; ++i) Col_L[i] = 0; + (*this).Set_Row_L_given_Col_L(); + } + else if (Nrows < Ncols) { + Row_L[0] = (idnr - 1)/Nrows + 1; + for (int i = 1; i < idnr - Nrows * (Row_L[0] - 1); ++i) Row_L[i] = 1; + for (int i = idnr - Nrows * (Row_L[0] - 1); i < Nrows; ++i) Row_L[i] = 0; + (*this).Set_Col_L_given_Row_L(); + } + } + + else { + + Young_Tableau Residual1(ndiag, Ncols - ndiag); + Young_Tableau Residual2(ndiag, Nrows - ndiag); + + Residual2.Set_to_id(residual_id/(Residual1.maxid + 1LL)); + residual_id -= (Residual1.maxid + 1LL) * Residual2.id; + Residual1.Set_to_id(residual_id); + + for (int i = 0; i < ndiag; ++i) Row_L[i] = ndiag + Residual1.Row_L[i]; + for (int i = 0; i < Residual2.Ncols; ++i) Row_L[ndiag + i] = Residual2.Col_L[i]; + + (*this).Set_Col_L_given_Row_L(); + + } + } + } // else if (option == 1) + + else if (option == 2) { + + Compute_Map (idnr); // make sure the state map is computed + (*this).Set_to_id (map[idnr], 0); + + id = idnr; + + } + + else JSCerror("Wrong option for Tableau id in Set_to_id"); + + return(*this); + } + + Young_Tableau& Young_Tableau::Set_Row_L (Vect_INT& Row_Lengths) // set row lengths to elements of given vector + { + if (Row_Lengths.size() != Nrows) JSCerror("Vector of incompatible dimension used to initialize Young Tableau."); + + for (int i = 0; i < Row_Lengths.size() - 1; ++i) if (Row_Lengths[i] < Row_Lengths[i+1]) JSCerror("Vector is not a proper Young tableau."); + + for (int i = 0; i < Nrows; ++i) Row_L[i] = Row_Lengths[i]; + (*this).Set_Col_L_given_Row_L(); + + return(*this); + } + + Young_Tableau& Young_Tableau::Set_Col_L_given_Row_L () // sets the Col_L array self-consistently + { + int col_l = 0; + for (int i = 0; i < Ncols; ++i) { + col_l = 0; + while ((Row_L[col_l] > i) && (col_l < Nrows)) ++col_l; + Col_L[i] = col_l; + } + return *this; + } + + Young_Tableau& Young_Tableau::Set_Row_L_given_Col_L () // sets the Col_L array self-consistently + { + int row_l = 0; + for (int i = 0; i < Nrows; ++i) { + row_l = 0; + while ((Col_L[row_l] > i) && (row_l < Ncols)) ++row_l; + Row_L[i] = row_l; + } + return *this; + } + + Young_Tableau& Young_Tableau::Print () // couts the tableau + { + cout << endl; + for (int i = 0; i < Nrows; ++i) { + if (Row_L[i] > 0) { + for (int j = 0; j < Row_L[i]; ++j) cout << "X"; + cout << endl; + } + } + cout << endl; + + return(*this); + } + + std::ostream& operator<< (std::ostream& s, const Young_Tableau& tableau) + { + s << endl; + + for (int i = 0; i < tableau.Nrows; ++i) { + if (tableau.Row_L[i] > 0) { + for (int j = 0; j < tableau.Row_L[i]; ++j) s << "X"; + s << endl; + } + } + s << endl; + + return(s); + } + + bool Young_Tableau::Lower_Row (int i) + { + // Removes a box from row i. Returns success boolean. + // Recomputes id. + + if (id == 0LL || Nrows == 0 || Ncols == 0) return(false); // Tableau is empty + if (i < 0 || i >= Nrows) JSCerror("Trying to Lower_Row of out of bounds index."); + if (Row_L[i] < 1 || (i < Nrows - 1 && Row_L[i+1] == Row_L[i])) // can't lower Row_L, breaks tableau rules + return(false); + else { + Row_L[i] -= 1; + Set_Col_L_given_Row_L(); + Compute_id(); + return(true); + } + return(false); + } + + bool Young_Tableau::Raise_Row (int i) + { + // Adds a box to row i. Returns success boolean. + // Recomputes id. + + if (Nrows == 0 || Ncols == 0) return(false); // Tableau is empty + if (i < 0 || i >= Nrows) JSCerror("Trying to Raise_Row of out of bounds index."); + if (Row_L[i] == Ncols || (i > 1 && Row_L[i-1] == Row_L[i])) // can't raise Row_L, breaks tableau rules + return(false); + else { + Row_L[i] += 1; + Set_Col_L_given_Row_L(); + Compute_id(); + return(true); + } + return(false); + } + + + bool Young_Tableau::Raise_Lowest_Nonzero_Row() + { + // adds a box to the lowest nonzero length Row, recomputes id, returns true if tableau has changed + + //cout << "Check before: "; (*this).Print(); + + if (id == 0LL || Nrows == 0 || Ncols == 0) return(false); // Tableau is empty + + // otherwise find the lowest nonzero row: + int iln0r = Nrows - 1; + while (Row_L[iln0r] == 0 && iln0r >= 0) iln0r--; + + if (iln0r < 0) JSCerror("id wrongly set in Young_Tableau (Raise_Lowest_Nonzero_Row)."); + // This should not happen, since if iln0r == -1, id should be 0. + + else if (iln0r == 0 && Row_L[0] < Ncols || iln0r > 0 && Row_L[iln0r - 1] > Row_L[iln0r]) { // there is space for at least one more box ! + Row_L[iln0r] += 1; + Set_Col_L_given_Row_L(); + Compute_id(); + + //cout << "Check after: iln0r = " << iln0r; (*this).Print(); + + return(true); + } + + return(false); + } + + bool Young_Tableau::Raise_Next_to_Lowest_Nonzero_Row() + { + // same thing, but for Row under lowest nonzero length one. + + // Important: allow raising first row if tableau is empty. + + //cout << "Check before: "; (*this).Print(); + + if (Ncols == 0 || Nrows == 0) return(false); // no space ! + + // Find index of lowest nonzero row: can be -1 if Tableau is empty + int iln0r = Nrows - 1; + while (Row_L[iln0r] == 0 && iln0r >= 0) iln0r--; + + //cout << "iln0r = " << iln0r << "\t" << Row_L[iln0r] << "\t" << Row_L[iln0r + 1] << endl; + + //if (iln0r == Nrows - 1) return(false); // no row under that one; allow raising of row 0 + + if (iln0r == -1 && Row_L[0] < Ncols || iln0r >= 0 && iln0r < Nrows - 1 && Row_L[iln0r] > Row_L[iln0r + 1]) { + // there is space for at least one more box ! + Row_L[iln0r + 1] += 1; + Set_Col_L_given_Row_L(); + Compute_id(); + + //cout << "Check after: iln0r = " << iln0r; (*this).Print(); + + return(true); + } + + return(false); + } + + bool Young_Tableau::Move_Box_from_Col_to_Col (int ifrom, int ito) + { + // Moves a box from column ifrom to column ito. Recomputes id. + // If fails: returns false, leaves Tableau unchanged. + + if (!(ifrom >= 0 && ifrom < Ncols && ito >= 0 && ito < Ncols)) + return(false); + + else { // try it out + int* Col_L_check = new int[Ncols]; + for (int i = 0; i < Ncols; ++i) Col_L_check[i] = Col_L[i]; + Col_L_check[ifrom] -= 1; + Col_L_check[ito] += 1; + if (Col_L_check[ifrom] < 0 || Col_L_check[ito] > Nrows) return(false); + for (int i = 0; i < Ncols - 1; ++i) + if (Col_L_check[i] < Col_L_check[i+1]) return(false); + // If we're here, it works. + Col_L[ifrom] -= 1; + Col_L[ito] += 1; + Set_Row_L_given_Col_L(); + Compute_id(); + } + return(true); + } + + Vect Young_Tableau::Descendents (int fixed_Nboxes) + { + // Produces a vector of Young_Tableau which are the descendents of + // the (*this) object. If the number of boxes is not fixed, there + // are up to 2 decendents. If the number of boxes is fixed, there + // can be many more. + + int ndesc = 0; + + if (!fixed_Nboxes) { + // Here, we can descend by adding a box to the lowest nonzero row, + // to to the empty one immediately below the lowest nonzero row: + + Young_Tableau Tableau_check1 = (*this); + bool check1 = (Tableau_check1.Raise_Lowest_Nonzero_Row()); + if (check1) ndesc++; + Young_Tableau Tableau_check2 = (*this); + bool check2 = (Tableau_check2.Raise_Next_to_Lowest_Nonzero_Row()); + if (check2) ndesc++; + + Vect Tableau_desc(ndesc); + if (check1) Tableau_desc[0] = Tableau_check1; + if (check2) Tableau_desc[check1] = Tableau_check2; + + return(Tableau_desc); + } + + else if (fixed_Nboxes) { + // Here, we can promote a box from the right to the left, + // provided it is in or to the left of col_bdry, where col_bdry + // is the column index of the leftmost column having a box which + // isn't `shadowed' when superimposing the minimal tableau having this + // number of boxes (all boxes in highest possible rows) with the + // actual tableau. + + // Define the original lowest tableau: + Young_Tableau Tableau_zero(*this); + int Nboxes_tot = 0; + for (int i = 0; i < Nrows; ++i) Nboxes_tot += Row_L[i]; + int Nboxes_above = 0; + for (int level = 0; level < Nrows; ++level) { + Tableau_zero.Row_L[level] = JSC::min(Nboxes_tot - Nboxes_above, Ncols); + Nboxes_above += Tableau_zero.Row_L[level]; + } + Tableau_zero.Set_Col_L_given_Row_L(); + + int level_bdry = 0; + for (int level = 0; level < Nrows; ++level) + if (Row_L[level] < Tableau_zero.Row_L[level]) level_bdry = level; + int right_bdry = JSC::min(Row_L[level_bdry] + 1, Ncols - 1); + // We can now displace a box from right to left, starting from + // a column with index <= right_bdry. + + int left_bdry = 0; + for (int level = 0; level < Ncols; ++level) + if (Col_L[level] > Tableau_zero.Col_L[level]) left_bdry = level; + // We can put a box into a column with index >= left_bdry + + // Now do the descendents: + Vect Tableau_desc_init ((right_bdry - left_bdry) * (right_bdry - left_bdry)); + Young_Tableau Tableau_ref = (*this); + Young_Tableau Tableau_check = (*this); + for (int ifrom = right_bdry; ifrom >= left_bdry + 1; --ifrom) + for (int ito = ifrom - 1; ito >= left_bdry; --ito) { + if (Tableau_check.Move_Box_from_Col_to_Col (ifrom, ito)) { + Tableau_desc_init[ndesc++] = Tableau_check; + Tableau_check = Tableau_ref; + } + } + + Vect Tableau_desc (ndesc); + for (int i = 0; i < ndesc; ++i) Tableau_desc[i] = Tableau_desc_init[i]; + + return(Tableau_desc); + + } // if (fixed_Nboxes) + + return(Vect (0)); + } + + Vect Young_Tableau::Descendents_Boosted_State (int fixed_Nboxes) + { + // Produces a vector of Young_Tableau which are the descendents of + // the (*this) object. If the number of boxes is not fixed, there + // are up to 2 decendents. If the number of boxes is fixed, there + // can be many more. + + // IMPORTANT ASSUMPTIONS: + // (*this) is (or is descended from) a boosted state. The only + // descendents considered here are thus those for which the raised + // box is the highest still occupied box of the originally boosted state. + + //cout << "Tableau in Desc_Boosted: " << (*this) << endl; + + int ndesc = 0; + + // Is tableau non-empty ? + if (Nrows <= 0 || Ncols <= 0) return(Vect (0)); + int Nboxes = 0; + for (int i = 0; i < Nrows; ++i) Nboxes += Row_L[i]; + + if (Nboxes == 0) return(Vect (0)); + + // Look for the level with the highest as yet unraised box of initial + // boosted state: + + int level_from = 0; + if (id == maxid) level_from = Nrows - 1; // full tableau treated here + + while (Row_L[level_from] == Ncols) level_from++; // necessarily is a row length < Ncols + + if (Row_L[level_from] == 0) level_from--; // if row empty or beyond limit, go back one + + if (!fixed_Nboxes) { + // The convention here is that we *remove* the highest yet unraised box only + + //cout << "Removing box from " << (*this) << " with id " << (*this).id << endl; + Young_Tableau descendent_attempt = (*this); + if (descendent_attempt.Lower_Row(level_from)) ndesc = 1; + //cout << "Obtained: " << descendent_attempt << " with id " << descendent_attempt.id << endl; + + //if (ndesc > 0) { + if (ndesc == 1) { + Vect Tableau_desc(ndesc); + //if (ndesc == 1) Tableau_desc[0] = descendent_attempt; + Tableau_desc[0] = descendent_attempt; + return(Tableau_desc); + } + else if (ndesc != 0) JSCerror("There should be either 0 or 1 descendents in Descended_Boosted_State with fixed_iK == true."); + + } // if (!fixed_Nboxes) + + else if (fixed_Nboxes) { + // We here move the same highest unraised box as for !fixed_Nboxes, + // except that we put higher up in this current tableau: + + // We already know from which level we take the box. + // We put it back either on the lowest level, or next: + + Young_Tableau Tableau_ref = (*this); + Young_Tableau Tableau_check1 = (*this); + bool check1 = (Tableau_check1.Lower_Row(level_from) && Tableau_check1.Raise_Lowest_Nonzero_Row()); + if (check1 && Tableau_check1.Row_L[level_from] == Tableau_ref.Row_L[level_from] - 1) ndesc++; // to make sure we don't Raise the one we've just removed + Young_Tableau Tableau_check2 = (*this); + bool check2 = (Tableau_check2.Lower_Row(level_from) && Tableau_check2.Raise_Next_to_Lowest_Nonzero_Row()); + if (check2 && Tableau_check2.Row_L[level_from] == Tableau_ref.Row_L[level_from] - 1) ndesc++; // to make sure we don't Raise the one we've just removed + + if (ndesc > 0) { + Vect Tableau_desc(ndesc); + if (check1) Tableau_desc[0] = Tableau_check1; + if (check2) Tableau_desc[ndesc - 1] = Tableau_check2; // only up to 2 descendents + return(Tableau_desc); + } + + } // if (fixed_Nboxes) + + return(Vect (0)); + } + + int Young_Tableau::Add_Boxes_From_Lowest (int Nboxes) + { + // tries to add Nboxes to Tableau, returns number of boxes added. + if (Ncols == 0 || Nrows == 0) return(0); // can't do anything ! + + //cout << "Requesting Nboxes " << Nboxes << " in tableau." << endl; + + int Nboxes_added = 0; + int previous_Row_L = 0; + for (int working_level = 0; working_level < Nrows; ++working_level) { + previous_Row_L = Row_L[working_level]; + Row_L[working_level] = JSC::min(previous_Row_L + Nboxes - Nboxes_added, Ncols); + Nboxes_added += Row_L[working_level] - previous_Row_L; + if (Nboxes_added == Nboxes) break; + } + Set_Col_L_given_Row_L(); + Compute_id(); + + return(Nboxes_added); + } + +} // namespace JSC