+ Start a Discussion

Debug Log Max Size Reached Even with filters set to NONE

I have tried everything:


Developer Console filters set to Error for all except Debug for Profiling

System monitoring everything set to none except Profiling

Apex Class, override and everything set to none excep profiling.



I still get max log sixed reached and never get to any profiling information. Here is a sample:


22:45:46.032 (32459000)|EXECUTION_STARTED
22:45:46.032 (32499000)|CODE_UNIT_STARTED|[EXTERNAL]|066Z00000004uBF|VF: /apex/pv_chem_bgreport_test
22:45:46.587 (587183000)|CODE_UNIT_STARTED|[EXTERNAL]|01pF0000002f6RI|PV_CHEM_BGReportController <init>
22:45:50.894 (4894671000)|CODE_UNIT_FINISHED|PV_CHEM_BGReportController <init>
22:45:50.894 (4894751000)|CODE_UNIT_STARTED|[EXTERNAL]|01pF0000002f6RI|PV_CHEM_BGReportController get(sParticipants)
22:45:50.894 (4894778000)|CODE_UNIT_STARTED|[EXTERNAL]|01pF0000002f6RI|sParticipants
22:45:50.894 (4894794000)|CODE_UNIT_FINISHED|sParticipants

 I thought all of this would be gone. Is that not the case? If so I cannot get around it as it displays ALL the values passed into the methods and there is a lot of data there.



Here is an example of one line with all the data being passed in showing


22:45:51.052 (5052874000)|CODE_UNIT_STARTED|[EXTERNAL]|PV_DemingRegAnalysisController set(objpv,PVReportObjects_Copy:[PTSampleType={c=venipuncture, i=skin puncture}, PVID=a2CF00000004zGqMAI, Panel=null, allPVData=(pvData&colon;[Name=PTINR, Panel=PT, SampleID=19, Units=INR, isOutlier=null, x1=2.7, x2=2.7, y1=2.2, y2=2], pvData&colon;[Name=PTINR, Panel=PT, SampleID=17, Units=INR, isOutlier=null, x1=1.2, x2=1.2, y1=1.1, y2=1.1], pvData&colon;[Name=PTINR, Panel=PT, SampleID=18, Units=INR, isOutlier=null, x1=3.7, x2=3.7, y1=2.8, y2=3.1], pvData&colon;[Name=PTINR, Panel=PT, SampleID=15, Units=INR, isOutlier=null, x1=1.7, x2=1.7, y1=1.5, y2=1.5], pvData&colon;[Name=PTINR, Panel=PT, SampleID=16, Units=INR, isOutlier=null, x1=1.9, x2=1.9, y1=2.1, y2=2], pvData&colon;[Name=PTINR, Panel=PT, SampleID=13, Units=INR, isOutlier=null, x1=1.4, x2=1.4, y1=1.5, y2=1.6], pvData&colon;[Name=PTINR, Panel=PT, SampleID=14, Units=INR, isOutlier=null, x1=1.5, x2=1.5, y1=1.4, y2=1.5], pvData&colon;[Name=PTINR, Panel=PT, SampleID=11, Units=INR, isOutlier=null, x1=2.1, x2=2.1, y1=2, y2=2.2], pvData&colon;[Name=PTINR, Panel=PT, SampleID=12, Units=INR, isOutlier=null, x1=2.9, x2=2.9, y1=3, y2=2.6], pvData&colon;[Name=PTINR, Panel=PT, SampleID=21, Units=INR, isOutlier=null, x1=3.1, x2=3.1, y1=2.4, y2=2.1], ...), anaStatPrecision={actwbt={CV=2, Diff=2, General=0, Intercept=1, Mean=1, R=2, Range=0, SD=2, Slope=1, Sxx=2, ...}, angap={CV=2, Diff=2, 

 This continues on and wraps for another 100 or so lines all for one line in the debug log where the filter is set to none.


Anyone have any ideas?


I have used below techniques in case if the debug log max size is reached.
1) If I want to the value of the variable. I used to use system.assert and compare the field(s) which you are expecting there could be issue and once the assert fails you can see the actual and expected value of that field.
2) If it is a Visualforce page and there is problem in the controller class. You can check the value of the field by including it in the outputtext field and check the value in front end.