Difference: Brenoorzari (23 vs. 24)

Revision 242019-08-29 - brenoorzari

Line: 1 to 1


Line: 270 to 270
3500 0.0024
4000 0.00083


First of all, Thiago said about writing a paper!!! I've already checked the works that cited the "Hunting the dark Higgs" (arXiv:1701.08780) paper, and must talk to Thiago about it.

The macro punzi_eff_bg_separated.C is being used to analyze cuts in the fat jet mass for the m_{d_{H}} = 70 GeV process. I'll do that for the other masses as well. New macros that are called punzi_eff_metcut.C, punzi_eff_etacut.C, punzi_eff_ptcut.C, and punzi_eff_lepcut.C were created to analyze every other cut. New cuts can be thought of at any time!!!

The macro V9_terceiromais_dark_Higgs.C was used at the processes with different m_{d_{H}}. An interesting thing happened at the fat jet ivariant mass from those processes: it had its peak displaced by about 10% for each mass (by comparing the m_{d_{H}} set mass with the m_{J} fitted mass). The graph from the comparison of those masses is in the image below:

  • mass_comp_ALL.png:

I've talked to Pedro and we concluded that the fat jet mass resolution is getting worse since the greater the m_{d_{H}} set, more non-collinear are the b's. To check that, I've tightened the MET cut (from 500 to 700), and we saw that the resolution got a bit better. A study about the efficiency of the cut may help with this problem. Also, this seems to be a characteristic of the jets and fat jets invariant mass resolution (the difference of m_{J} and m_{d_{H}} scales with the mass).

The macro V9_terceiromais_dark_Higgs.C was also used at the processes with different m_{Z'} and the results were just as I've expected.


All the macros are working perfectly. The macro punzi_eff_lepcut.C needed to have the number of events exchanged, since the restriction in the lepton p_T made the cut looser (and the relative efficiencies were getting higher than 1). I must talk to Thiago about this.

I'm using the fat jet pt macro in all the other dark Higgs masses to see what happens to the efficiency. Luckily, it will change something. I'll also use the macro of the fat jet mass to analyze the events.


I'm running all the macros again with the Z' mass variations, only to check what is happening. I still need to plot signal and background MET one over another to check if the MET efficiency is alright. If there is something missing, I can generate events with a looser cut in MET and merge it with the ones I already have.


After running the macro to calculate the efficiencies of the MET cut, no important result was found, in the sense that the significance only goes down. I'll check why that is happening. Also, me and Thiago saw that some of the processes generated by the MadGraph have a Higgs or a dark Higgs mediating the production of two b's and two DM particles. I'll see the lagrangian implemented inside the UFO files.

I've created a macro to plot the MET and the fat jet p_T in the same histogram to see what is happening with the Punzi significance. Thiago said that is important to have the BG separated being compared with the signal, but also the BG all together being compared with the signal. I've only rescaled the histogram by each of the processes cross sections. Later, I can normalize everything to see what happens. Thiago have showed me a variable that might be interesting to plot and cut it in some interesting region: \sqrt{2 * p_{T}^{J} * MET * (1-cos(\Delta \phi (J,MET)))}.

At tuesdays and wednesdays I'll start participating in the dark Higgs technical chat (tuesdays) and the monojet meeting (wednesdays). This will be very nice to learn a lot of new stuff with people that work in the same stuff that I'm doing right now.

I've generated histograms of the transverse mass, and it really seems that it can be cut to increase the significance of the search!!!


Yesterday I've listened to the monojet meeting, and they were talking about some discrepancies in some monojet samples cross sections, that could be caused by the use of a wrong PDF.

Today, I've created a macro to analyze the Punzi significance of cuts in the transverse mass, and it really seems that awesome results will come out. I'll show this in today's SPRACE physics meeting. I think I should run the macros to the other masses as well.

The results from the Punzi significance for cuts in the transverse mass show a very small improvement in the significance for one of the cuts, and I'll show it in todays meeting. It's not a great gain, but it's something.


At the 01/08 meeting, Pedro sugested that I should cut the M_{T} variable in a window around the Z' mass. Maybe I'll heva some gain in efficiency here. I don't remember exactly when, but I have written a macro to plot the cross sections of the processes for different m_{d_{H}} or m_{Z'}. It' named plot_dh_xsec_wo_cuts.C (since it only has the MadGraph+Pythia+Delphes generation cuts), and the files created were xsec_mdh_var_nocuts.pdf and xsec_mdh_var_nocuts.pdf.

I'll begin to create folders for each different macro that I'll use. All the files related to a test of that macro will be in subfolders inside the "master" one related to that macro. I hope that all my files get more organized in this way. The folder name will be exactly like the macro name.


Yesterday I've showed to Thiago the results of the Punzi significance for the transverse mass cuts for the file with m_{d_{H}} = 70 GeV and m_{Z'} = 1100 GeV, and it wasn't very satisfying. Thiago asked me to do the same with a bigger m_{Z'} file, and I did for the 2000 GeV one, which got us even more confused. I'm using the LHE files to calculate the p_{T} of the sum of the four-momentum of the DM particles, and it's exactly like the MET in the root file corresponding to this mass.

The macro name is read_fourvector.C, and there is a folder with the same name, with all the files/canvases of this macro.


I've presented a summary of the CWR paper yesterday at the meeting, and already sent my comments/doubts to Pedro.

Thiago helped me to realise that there was an error in the macro read_fourvector.C. I've fixed it, and inside the folder read_fourvector, there is another folder with the name macro_certa, where all the right plots are, including a figure that compare the MET histograms for different values of m_{Z'}.

Today me and Thiago had a talk about what should I do in the next few weeks, and these are my tasks:

*Build a table with all the number of events; *Build a cut flow for some (all) of the m_{D_{H}} values; *Learn how to use ROOT TLimit or TRolke (or both);


I've modified the macro punzi_eff_bg_separated to find the best window around the best center value (that I got through the highest Punzi significance). The files are inside the folder best_punzi_center_varying_binsize in the folder named after the macro. Now, I'll modify it to find the number of events of signal and BG after all the cuts, in the region of highest Punzi significance for the center and the window, and it will be named punzi_eff_table_cutflow.C.

There were some errors in the tables, and I'll se what is happening tomorrow. But it seems that I'm at the right way.


The first table is below. The number of events is normalized to a luminosity of 140 fb^{-1}

m_{d_{H}} [GeV] N_{S} N_{B} m_{J}^{center} [GeV] m_{J}^{window} [GeV]
50 77.8346 59.5086 50 20
60 101.74 47.6723 58 16
70 136.038 68.2943 66 20
80 166.037 122.624 74 28
90 172.762 136.005 82 28
100 175.813 140.486 98 36
110 174.741 131.641 110 40
120 166.938 105.836 112 36
130 122.976 87.6493 122 32
140 104.722 136.891 130 44
150 57.0858 165.731 146 44
160 18.6357 271.383 150 72
170 3.43249 250.224 160 64
180 2.07056 218.363 166 56

Some examples of cutflow are

m_{d_{H}} = 50(50\pm10) [GeV] N_{S} N_{B} s_{Punzi} N_{S}/(sqrt{N_{S}+N_{B}}) novo BG s_{Punzi}
MET 1163.78 7383.93 0.00031072 12.5877 0.000298948
Lep 1163.78 7383.93 0.00031072 12.5877 0.000298948
Fat jet p_{T} 1159.02 7374.31 0.000309648 12.5468 0.000297998
Fat jet eta 1152.9 7324.61 0.000309039 12.5216 0.000297501
Fat jet mass 77.8346 59.5086 0.000197912 6.64154 0.000187386

m_{d_{H}} = 50(50\pm10) [GeV] N_{S} N_{B} s_{Punzi} N_{S}/(sqrt{N_{S}+N_{B}})
MET 1163.78 7383.93 0.00031072 12.5877
Lep 1163.78 7383.93 0.00381325 12.5877
Fat jet p_{T} 1159.02 7374.31 0.00380012 12.5468
Fat jet eta 1152.9 7324.61 0.00380592 12.5216
Fat jet mass 77.8346 59.5086 0.00275026 6.64154

m_{d_{H}} = 110(110\pm20) [GeV] N_{S} N_{B} s_{Punzi} N_{S}/(sqrt{N_{S}+N_{B}}) novo BG s_{Punzi}
MET 917.668 7383.93 0.000450836 10.0718 0.000433755
Lep 917.668 7383.93 0.000450836 10.0718 0.000433755
Fat jet p_{T} 912.289 7374.31 0.00044848 10.0218 0.000431608
Fat jet eta 905.982 7324.61 0.000446863 9.98628 0.000430179
Fat jet mass 174.741 131.641 0.000579978 9.98305 0.000546206

m_{d_{H}} = 110(110\pm20) [GeV] N_{S} N_{B} s_{Punzi} N_{S}/(sqrt{N_{S}+N_{B}})
MET 917.668 7383.93 0.000450836 10.0718
Lep 917.668 7383.93 0.00381325 10.0718
Fat jet p_{T} 912.289 7374.31 0.00379336 10.0218
Fat jet eta 905.982 7324.61 0.00379966 9.98628
Fat jet mass 174.741 131.641 0.00535427 9.98305

m_{d_{H}} = 180(166\pm28) [GeV] N_{S} N_{B} s_{Punzi} N_{S}/(sqrt{N_{S}+N_{B}}) novo BG s_{Punzi}
MET 9.26213 7383.93 0.000587774 0.10772 0.000565505
Lep 9.26213 7383.93 0.000587774 0.10772 0.000565505
Fat jet p_{T} 9.19162 7374.31 0.000583673 0.10697 0.000561715
Fat jet eta 9.14414 7324.61 0.000582592 0.106777 0.00056084
Fat jet mass 2.07056 218.363 0.000707119 0.13946 0.000698256

m_{d_{H}} = 180(166\pm28) [GeV] N_{S} N_{B} s_{Punzi} N_{S}/(sqrt{N_{S}+N_{B}})
MET 9.26213 7383.93 0.000587774 0.10772
Lep 9.26213 7383.93 0.00381325 0.10772
Fat jet p_{T} 9.19162 7374.31 0.00378668 0.10697
Fat jet eta 9.14414 7324.61 0.00380635 0.106777
Fat jet mass 2.07056 218.363 0.00491111 0.13946

I can't understand why the Punzi significance or the number of events statistical significance is getting lower after each cut.


Today I've been working on the suggestions everyone gave me at last thursday's SPRACE physics meeting. I'm still not sure about the Punzi significance getting lower, and I'll Thiago to show him these results. I've also changed the calculation of BG for the Punzi significance to the expression that Ana and Tulio are using that is the sum of the efficiency times the cross section times the luminosity of each BG separately, but it didn't changed much in the results. I'll talk to someone about this and check what is the right expression.

I've created a new macro called punzi_eff_table_cutflow_lepbef to check what happens if I do the lepton veto before the MET cut, and how this changes the Punzi significance.

I've created a macro to test the TLimit class from ROOT. It's named tlimit.C and it uses the ROOT file 140_fb_newbin_xsecfactor_ATLAS_kfactors_NLO_bgchange_mais_certo_ALL_bg_sig_70gev_withiso.root since it already have 140 fb^{-1} of luminosity and it has BG + Sig to test. I'm still very confused about this class, and what its outputs represents for the dark Higgs model. But I've only tested with one mass, and I must do it to the other masses as well.

I'm doing the poster for ENFPC, and I'll get some stuff from the previous poster. I might put something about the validation of the model (following the dark higgs paper) and some of these efficiencies/significances results.


In the folder NLO_mais_certo_ALL_backgrounds_combined_rightratio are the histograms that compare the fat jet mass of BG only, signal only and BG+sig for all the m_{d_{H}} values. I'll use those to test TLimit and TRolke. The files are named 140_fb_newbin_xsecfactor_ATLAS_kfactors_NLO_bgchange_mais_certo_ALL_bg_sig_"+m_{d_{H}}+"gev_withiso.root.

I've created a macro to test the TRolke ROOT class, that is named trolke_example.C, that is in its ROOT class reference. I'm still trying to figure out what to do with this, and what efficiencies should I use in my case. Also, what is the ratio between signal and BG regions.


Today I've talked to Thiago, and I'm testing the TLimit ROOT class only using some number of events. For now I've done for 3 events, 5 events and 10 events, in 8 different cases: (sig, bg, data) = (0,0,0); (n,0,0); (0,n,0); (0,0,n); (n,n,0); (n,0,n); (0,n,n); (n,n,n). Actually, I need to use 0.0000000001 events to represent 0, since I was getting some error messages.

I'll start doubling the last case data number of events, to see what happens (I'm expecting that CLs and CLsb will increase as n is higher).


I've tested the TLimit class, and it works reasonably good. Using it, I could set exclusion masses for d_{H} and Z'. The table for the number of events for each m_{Z'} is (the luminosity was set to 140 fb^{-1})

m_{Z'} [GeV] N_{S} N_{B} m_{J}^{center} [GeV] m_{J}^{window} [GeV]
500 228.074 68.2943 66 20
625 216.908 68.2943 66 20
1000 142.014 68.2943 66 20
1100 136.038 68.2943 66 20
1500 128.621 68.2943 66 20
1700 101.867 68.2943 66 20
2000 67.6939 68.2943 66 20
2500 28.3811 68.2943 66 20
3000 11.3235 68.2943 66 20
3500 4.58453 68.2943 66 20
4000 1.72912 68.2943 66 20

The graphs are below. Everything above the dashed lines is excluded with more than 95% CL.

  • cls_mdh.png:

  • cls_mzp.png:

In the first plot it was set m_{Z'} = 1100 GeV and m_{\chi} = 100 GeV. In the second plot it is m_{d_{H}} = 70 GeV and m_{\chi} = 100 GeV.



Line: 290 to 494
META FILEATTACHMENT attachment="bg_certo_comp.png" attr="" comment="" date="1562167368" name="bg_certo_comp.png" path="bg_certo_comp.png" size="49830" user="brenoorzari" version="1"
META FILEATTACHMENT attachment="comp_bg_NLO.png" attr="" comment="" date="1562258610" name="comp_bg_NLO.png" path="comp_bg_NLO.png" size="57438" user="brenoorzari" version="1"
META FILEATTACHMENT attachment="xsec_factor_bg.png" attr="" comment="" date="1562268796" name="xsec_factor_bg.png" path="xsec_factor_bg.png" size="40406" user="brenoorzari" version="1"
META FILEATTACHMENT attachment="mass_comp_ALL.png" attr="" comment="" date="1567116276" name="mass_comp_ALL.png" path="mass_comp_ALL.png" size="34514" user="brenoorzari" version="1"
META FILEATTACHMENT attachment="cls_mdh.png" attr="" comment="" date="1567116450" name="cls_mdh.png" path="cls_mdh.png" size="26017" user="brenoorzari" version="1"
META FILEATTACHMENT attachment="cls_mzp.png" attr="" comment="" date="1567116461" name="cls_mzp.png" path="cls_mzp.png" size="25445" user="brenoorzari" version="1"
This site is powered by the TWiki collaboration platform Powered by PerlCopyright © 2008-2021 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki? Send feedback

antalya escort bursa escort eskisehir escort istanbul escort izmir escort