1 |
diff -up jdk8/hotspot/src/share/vm/gc_implementation/concurrentMarkSweep/compactibleFreeListSpace.cpp.size_t jdk8/hotspot/src/share/vm/gc_implementation/concurrentMarkSweep/compactibleFreeListSpace.cpp |
2 |
--- jdk8/hotspot/src/share/vm/gc_implementation/concurrentMarkSweep/compactibleFreeListSpace.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
3 |
+++ jdk8/hotspot/src/share/vm/gc_implementation/concurrentMarkSweep/compactibleFreeListSpace.cpp 2014-05-16 13:09:21.350132543 -0400 |
4 |
@@ -2660,7 +2660,7 @@ void CFLS_LAB::get_from_global_pool(size |
5 |
if (ResizeOldPLAB && CMSOldPLABResizeQuicker) { |
6 |
size_t multiple = _num_blocks[word_sz]/(CMSOldPLABToleranceFactor*CMSOldPLABNumRefills*n_blks); |
7 |
n_blks += CMSOldPLABReactivityFactor*multiple*n_blks; |
8 |
- n_blks = MIN2(n_blks, CMSOldPLABMax); |
9 |
+ n_blks = MIN2(n_blks, (size_t)CMSOldPLABMax); |
10 |
} |
11 |
assert(n_blks > 0, "Error"); |
12 |
_cfls->par_get_chunk_of_blocks(word_sz, n_blks, fl); |
13 |
diff -up jdk8/hotspot/src/share/vm/gc_implementation/concurrentMarkSweep/concurrentMarkSweepGeneration.cpp.size_t jdk8/hotspot/src/share/vm/gc_implementation/concurrentMarkSweep/concurrentMarkSweepGeneration.cpp |
14 |
--- jdk8/hotspot/src/share/vm/gc_implementation/concurrentMarkSweep/concurrentMarkSweepGeneration.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
15 |
+++ jdk8/hotspot/src/share/vm/gc_implementation/concurrentMarkSweep/concurrentMarkSweepGeneration.cpp 2014-05-16 13:09:21.350132543 -0400 |
16 |
@@ -949,7 +949,7 @@ void ConcurrentMarkSweepGeneration::comp |
17 |
if (free_percentage < desired_free_percentage) { |
18 |
size_t desired_capacity = (size_t)(used() / ((double) 1 - desired_free_percentage)); |
19 |
assert(desired_capacity >= capacity(), "invalid expansion size"); |
20 |
- size_t expand_bytes = MAX2(desired_capacity - capacity(), MinHeapDeltaBytes); |
21 |
+ size_t expand_bytes = MAX2(desired_capacity - capacity(), (size_t)MinHeapDeltaBytes); |
22 |
if (PrintGCDetails && Verbose) { |
23 |
size_t desired_capacity = (size_t)(used() / ((double) 1 - desired_free_percentage)); |
24 |
gclog_or_tty->print_cr("\nFrom compute_new_size: "); |
25 |
@@ -6555,7 +6555,7 @@ void CMSCollector::reset(bool asynch) { |
26 |
HeapWord* curAddr = _markBitMap.startWord(); |
27 |
while (curAddr < _markBitMap.endWord()) { |
28 |
size_t remaining = pointer_delta(_markBitMap.endWord(), curAddr); |
29 |
- MemRegion chunk(curAddr, MIN2(CMSBitMapYieldQuantum, remaining)); |
30 |
+ MemRegion chunk(curAddr, MIN2((size_t)CMSBitMapYieldQuantum, remaining)); |
31 |
_markBitMap.clear_large_range(chunk); |
32 |
if (ConcurrentMarkSweepThread::should_yield() && |
33 |
!foregroundGCIsActive() && |
34 |
@@ -6854,7 +6854,7 @@ void CMSMarkStack::expand() { |
35 |
return; |
36 |
} |
37 |
// Double capacity if possible |
38 |
- size_t new_capacity = MIN2(_capacity*2, MarkStackSizeMax); |
39 |
+ size_t new_capacity = MIN2(_capacity*2, (size_t)MarkStackSizeMax); |
40 |
// Do not give up existing stack until we have managed to |
41 |
// get the double capacity that we desired. |
42 |
ReservedSpace rs(ReservedSpace::allocation_align_size_up( |
43 |
diff -up jdk8/hotspot/src/share/vm/gc_implementation/g1/concurrentMark.cpp.size_t jdk8/hotspot/src/share/vm/gc_implementation/g1/concurrentMark.cpp |
44 |
--- jdk8/hotspot/src/share/vm/gc_implementation/g1/concurrentMark.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
45 |
+++ jdk8/hotspot/src/share/vm/gc_implementation/g1/concurrentMark.cpp 2014-05-16 13:09:21.350132543 -0400 |
46 |
@@ -3745,7 +3745,7 @@ void CMTask::drain_local_queue(bool part |
47 |
// of things to do) or totally (at the very end). |
48 |
size_t target_size; |
49 |
if (partially) { |
50 |
- target_size = MIN2((size_t)_task_queue->max_elems()/3, GCDrainStackTargetSize); |
51 |
+ target_size = MIN2((size_t)(_task_queue->max_elems()/3), (size_t) GCDrainStackTargetSize); |
52 |
} else { |
53 |
target_size = 0; |
54 |
} |
55 |
@@ -4583,7 +4583,7 @@ size_t G1PrintRegionLivenessInfoClosure: |
56 |
// The > 0 check is to deal with the prev and next live bytes which |
57 |
// could be 0. |
58 |
if (*hum_bytes > 0) { |
59 |
- bytes = MIN2(HeapRegion::GrainBytes, *hum_bytes); |
60 |
+ bytes = MIN2(HeapRegion::GrainBytes, (size_t)*hum_bytes); |
61 |
*hum_bytes -= bytes; |
62 |
} |
63 |
return bytes; |
64 |
diff -up jdk8/hotspot/src/share/vm/gc_implementation/g1/g1CollectedHeap.cpp.size_t jdk8/hotspot/src/share/vm/gc_implementation/g1/g1CollectedHeap.cpp |
65 |
--- jdk8/hotspot/src/share/vm/gc_implementation/g1/g1CollectedHeap.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
66 |
+++ jdk8/hotspot/src/share/vm/gc_implementation/g1/g1CollectedHeap.cpp 2014-05-16 13:09:21.360132543 -0400 |
67 |
@@ -1742,7 +1742,7 @@ HeapWord* G1CollectedHeap::expand_and_al |
68 |
|
69 |
verify_region_sets_optional(); |
70 |
|
71 |
- size_t expand_bytes = MAX2(word_size * HeapWordSize, MinHeapDeltaBytes); |
72 |
+ size_t expand_bytes = MAX2(word_size * HeapWordSize, (size_t)MinHeapDeltaBytes); |
73 |
ergo_verbose1(ErgoHeapSizing, |
74 |
"attempt heap expansion", |
75 |
ergo_format_reason("allocation request failed") |
76 |
diff -up jdk8/hotspot/src/share/vm/gc_implementation/g1/heapRegion.cpp.size_t jdk8/hotspot/src/share/vm/gc_implementation/g1/heapRegion.cpp |
77 |
--- jdk8/hotspot/src/share/vm/gc_implementation/g1/heapRegion.cpp.size_t 2014-05-16 13:35:15.150132543 -0400 |
78 |
+++ jdk8/hotspot/src/share/vm/gc_implementation/g1/heapRegion.cpp 2014-05-16 13:35:48.090132543 -0400 |
79 |
@@ -158,7 +158,7 @@ void HeapRegion::setup_heap_region_size( |
80 |
if (FLAG_IS_DEFAULT(G1HeapRegionSize)) { |
81 |
size_t average_heap_size = (initial_heap_size + max_heap_size) / 2; |
82 |
region_size = MAX2(average_heap_size / TARGET_REGION_NUMBER, |
83 |
- (uintx) MIN_REGION_SIZE); |
84 |
+ (size_t) MIN_REGION_SIZE); |
85 |
} |
86 |
|
87 |
int region_size_log = log2_long((jlong) region_size); |
88 |
diff -up jdk8/hotspot/src/share/vm/gc_implementation/g1/heapRegionRemSet.cpp.size_t jdk8/hotspot/src/share/vm/gc_implementation/g1/heapRegionRemSet.cpp |
89 |
--- jdk8/hotspot/src/share/vm/gc_implementation/g1/heapRegionRemSet.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
90 |
+++ jdk8/hotspot/src/share/vm/gc_implementation/g1/heapRegionRemSet.cpp 2014-05-16 13:09:21.360132543 -0400 |
91 |
@@ -845,7 +845,7 @@ OtherRegionsTable::do_cleanup_work(HRRSC |
92 |
// This can be done by either mutator threads together with the |
93 |
// concurrent refinement threads or GC threads. |
94 |
int HeapRegionRemSet::num_par_rem_sets() { |
95 |
- return (int)MAX2(DirtyCardQueueSet::num_par_ids() + ConcurrentG1Refine::thread_num(), ParallelGCThreads); |
96 |
+ return (int)MAX2(DirtyCardQueueSet::num_par_ids() + ConcurrentG1Refine::thread_num(), (size_t)ParallelGCThreads); |
97 |
} |
98 |
|
99 |
HeapRegionRemSet::HeapRegionRemSet(G1BlockOffsetSharedArray* bosa, |
100 |
diff -up jdk8/hotspot/src/share/vm/gc_implementation/parallelScavenge/parMarkBitMap.cpp.size_t jdk8/hotspot/src/share/vm/gc_implementation/parallelScavenge/parMarkBitMap.cpp |
101 |
--- jdk8/hotspot/src/share/vm/gc_implementation/parallelScavenge/parMarkBitMap.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
102 |
+++ jdk8/hotspot/src/share/vm/gc_implementation/parallelScavenge/parMarkBitMap.cpp 2014-05-16 13:09:21.360132543 -0400 |
103 |
@@ -68,7 +68,7 @@ ParMarkBitMap::initialize(MemRegion cove |
104 |
if (_virtual_space != NULL && _virtual_space->expand_by(_reserved_byte_size)) { |
105 |
_region_start = covered_region.start(); |
106 |
_region_size = covered_region.word_size(); |
107 |
- idx_t* map = (idx_t*)_virtual_space->reserved_low_addr(); |
108 |
+ BitMap::bm_word_t* map = (BitMap::bm_word_t*)_virtual_space->reserved_low_addr(); |
109 |
_beg_bits.set_map(map); |
110 |
_beg_bits.set_size(bits / 2); |
111 |
_end_bits.set_map(map + words / 2); |
112 |
diff -up jdk8/hotspot/src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.cpp.size_t jdk8/hotspot/src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.cpp |
113 |
--- jdk8/hotspot/src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
114 |
+++ jdk8/hotspot/src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.cpp 2014-05-16 13:09:21.360132543 -0400 |
115 |
@@ -909,8 +909,8 @@ void PSParallelCompact::initialize_space |
116 |
void PSParallelCompact::initialize_dead_wood_limiter() |
117 |
{ |
118 |
const size_t max = 100; |
119 |
- _dwl_mean = double(MIN2(ParallelOldDeadWoodLimiterMean, max)) / 100.0; |
120 |
- _dwl_std_dev = double(MIN2(ParallelOldDeadWoodLimiterStdDev, max)) / 100.0; |
121 |
+ _dwl_mean = double(MIN2((size_t)ParallelOldDeadWoodLimiterMean, max)) / 100.0; |
122 |
+ _dwl_std_dev = double(MIN2((size_t)ParallelOldDeadWoodLimiterStdDev, max)) / 100.0; |
123 |
_dwl_first_term = 1.0 / (sqrt(2.0 * M_PI) * _dwl_std_dev); |
124 |
DEBUG_ONLY(_dwl_initialized = true;) |
125 |
_dwl_adjustment = normal_distribution(1.0); |
126 |
diff -up jdk8/hotspot/src/share/vm/gc_implementation/parNew/parNewGeneration.cpp.size_t jdk8/hotspot/src/share/vm/gc_implementation/parNew/parNewGeneration.cpp |
127 |
--- jdk8/hotspot/src/share/vm/gc_implementation/parNew/parNewGeneration.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
128 |
+++ jdk8/hotspot/src/share/vm/gc_implementation/parNew/parNewGeneration.cpp 2014-05-16 13:09:21.360132543 -0400 |
129 |
@@ -192,7 +192,7 @@ bool ParScanThreadState::take_from_overf |
130 |
const size_t num_overflow_elems = of_stack->size(); |
131 |
const size_t space_available = queue->max_elems() - queue->size(); |
132 |
const size_t num_take_elems = MIN3(space_available / 4, |
133 |
- ParGCDesiredObjsFromOverflowList, |
134 |
+ (size_t)ParGCDesiredObjsFromOverflowList, |
135 |
num_overflow_elems); |
136 |
// Transfer the most recent num_take_elems from the overflow |
137 |
// stack to our work queue. |
138 |
diff -up jdk8/hotspot/src/share/vm/memory/collectorPolicy.cpp.size_t jdk8/hotspot/src/share/vm/memory/collectorPolicy.cpp |
139 |
--- jdk8/hotspot/src/share/vm/memory/collectorPolicy.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
140 |
+++ jdk8/hotspot/src/share/vm/memory/collectorPolicy.cpp 2014-05-16 13:31:55.620132543 -0400 |
141 |
@@ -386,7 +386,7 @@ void TwoGenerationCollectorPolicy::initi |
142 |
uintx calculated_size = NewSize + OldSize; |
143 |
double shrink_factor = (double) MaxHeapSize / calculated_size; |
144 |
uintx smaller_new_size = align_size_down((uintx)(NewSize * shrink_factor), _gen_alignment); |
145 |
- FLAG_SET_ERGO(uintx, NewSize, MAX2(young_gen_size_lower_bound(), smaller_new_size)); |
146 |
+ FLAG_SET_ERGO(uintx, NewSize, MAX2(young_gen_size_lower_bound(), (size_t)smaller_new_size)); |
147 |
_initial_gen0_size = NewSize; |
148 |
|
149 |
// OldSize is already aligned because above we aligned MaxHeapSize to |
150 |
@@ -434,7 +434,7 @@ void GenCollectorPolicy::initialize_size |
151 |
// yield a size that is too small) and bound it by MaxNewSize above. |
152 |
// Ergonomics plays here by previously calculating the desired |
153 |
// NewSize and MaxNewSize. |
154 |
- max_new_size = MIN2(MAX2(max_new_size, NewSize), MaxNewSize); |
155 |
+ max_new_size = MIN2(MAX2(max_new_size, (size_t)NewSize), (size_t)MaxNewSize); |
156 |
} |
157 |
assert(max_new_size > 0, "All paths should set max_new_size"); |
158 |
|
159 |
@@ -461,16 +461,16 @@ void GenCollectorPolicy::initialize_size |
160 |
// generally small compared to the NewRatio calculation. |
161 |
_min_gen0_size = NewSize; |
162 |
desired_new_size = NewSize; |
163 |
- max_new_size = MAX2(max_new_size, NewSize); |
164 |
+ max_new_size = MAX2(max_new_size, (size_t)NewSize); |
165 |
} else { |
166 |
// For the case where NewSize is the default, use NewRatio |
167 |
// to size the minimum and initial generation sizes. |
168 |
// Use the default NewSize as the floor for these values. If |
169 |
// NewRatio is overly large, the resulting sizes can be too |
170 |
// small. |
171 |
- _min_gen0_size = MAX2(scale_by_NewRatio_aligned(_min_heap_byte_size), NewSize); |
172 |
+ _min_gen0_size = MAX2(scale_by_NewRatio_aligned(_min_heap_byte_size), (size_t)NewSize); |
173 |
desired_new_size = |
174 |
- MAX2(scale_by_NewRatio_aligned(_initial_heap_byte_size), NewSize); |
175 |
+ MAX2(scale_by_NewRatio_aligned(_initial_heap_byte_size), (size_t)NewSize); |
176 |
} |
177 |
|
178 |
assert(_min_gen0_size > 0, "Sanity check"); |
179 |
@@ -572,7 +572,7 @@ void TwoGenerationCollectorPolicy::initi |
180 |
} else { |
181 |
// It's been explicitly set on the command line. Use the |
182 |
// OldSize and then determine the consequences. |
183 |
- _min_gen1_size = MIN2(OldSize, _min_heap_byte_size - _min_gen0_size); |
184 |
+ _min_gen1_size = MIN2((size_t)OldSize, _min_heap_byte_size - _min_gen0_size); |
185 |
_initial_gen1_size = OldSize; |
186 |
|
187 |
// If the user has explicitly set an OldSize that is inconsistent |
188 |
diff -up jdk8/hotspot/src/share/vm/memory/metaspace.cpp.size_t jdk8/hotspot/src/share/vm/memory/metaspace.cpp |
189 |
--- jdk8/hotspot/src/share/vm/memory/metaspace.cpp.size_t 2014-05-16 13:40:21.870132543 -0400 |
190 |
+++ jdk8/hotspot/src/share/vm/memory/metaspace.cpp 2014-05-16 13:44:26.930132543 -0400 |
191 |
@@ -1424,7 +1424,7 @@ void MetaspaceGC::compute_new_size() { |
192 |
(size_t)MIN2(min_tmp, double(max_uintx)); |
193 |
// Don't shrink less than the initial generation size |
194 |
minimum_desired_capacity = MAX2(minimum_desired_capacity, |
195 |
- MetaspaceSize); |
196 |
+ (size_t)MetaspaceSize); |
197 |
|
198 |
if (PrintGCDetails && Verbose) { |
199 |
gclog_or_tty->print_cr("\nMetaspaceGC::compute_new_size: "); |
200 |
@@ -1477,7 +1477,7 @@ void MetaspaceGC::compute_new_size() { |
201 |
const double max_tmp = used_after_gc / minimum_used_percentage; |
202 |
size_t maximum_desired_capacity = (size_t)MIN2(max_tmp, double(max_uintx)); |
203 |
maximum_desired_capacity = MAX2(maximum_desired_capacity, |
204 |
- MetaspaceSize); |
205 |
+ (size_t)MetaspaceSize); |
206 |
if (PrintGCDetails && Verbose) { |
207 |
gclog_or_tty->print_cr(" " |
208 |
" maximum_free_percentage: %6.2f" |
209 |
@@ -3110,7 +3110,7 @@ void Metaspace::global_initialize() { |
210 |
// on the medium chunk list. The next chunk will be small and progress |
211 |
// from there. This size calculated by -version. |
212 |
_first_class_chunk_word_size = MIN2((size_t)MediumChunk*6, |
213 |
- (CompressedClassSpaceSize/BytesPerWord)*2); |
214 |
+ (size_t)(CompressedClassSpaceSize/BytesPerWord)*2); |
215 |
_first_class_chunk_word_size = align_word_size_up(_first_class_chunk_word_size); |
216 |
// Arbitrarily set the initial virtual space to a multiple |
217 |
// of the boot class loader size. |
218 |
diff -up jdk8/hotspot/src/share/vm/memory/threadLocalAllocBuffer.cpp.size_t jdk8/hotspot/src/share/vm/memory/threadLocalAllocBuffer.cpp |
219 |
--- jdk8/hotspot/src/share/vm/memory/threadLocalAllocBuffer.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
220 |
+++ jdk8/hotspot/src/share/vm/memory/threadLocalAllocBuffer.cpp 2014-05-16 13:09:21.360132543 -0400 |
221 |
@@ -232,7 +232,7 @@ size_t ThreadLocalAllocBuffer::initial_d |
222 |
size_t init_sz; |
223 |
|
224 |
if (TLABSize > 0) { |
225 |
- init_sz = MIN2(TLABSize / HeapWordSize, max_size()); |
226 |
+ init_sz = MIN2((size_t)(TLABSize / HeapWordSize), max_size()); |
227 |
} else if (global_stats() == NULL) { |
228 |
// Startup issue - main thread initialized before heap initialized. |
229 |
init_sz = min_size(); |
230 |
diff -up jdk8/hotspot/src/share/vm/oops/objArrayKlass.inline.hpp.size_t jdk8/hotspot/src/share/vm/oops/objArrayKlass.inline.hpp |
231 |
--- jdk8/hotspot/src/share/vm/oops/objArrayKlass.inline.hpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
232 |
+++ jdk8/hotspot/src/share/vm/oops/objArrayKlass.inline.hpp 2014-05-16 13:09:21.360132543 -0400 |
233 |
@@ -48,7 +48,7 @@ void ObjArrayKlass::objarray_follow_cont |
234 |
const size_t beg_index = size_t(index); |
235 |
assert(beg_index < len || len == 0, "index too large"); |
236 |
|
237 |
- const size_t stride = MIN2(len - beg_index, ObjArrayMarkingStride); |
238 |
+ const size_t stride = MIN2(len - beg_index, (size_t)ObjArrayMarkingStride); |
239 |
const size_t end_index = beg_index + stride; |
240 |
T* const base = (T*)a->base(); |
241 |
T* const beg = base + beg_index; |
242 |
@@ -82,7 +82,7 @@ void ObjArrayKlass::objarray_follow_cont |
243 |
const size_t beg_index = size_t(index); |
244 |
assert(beg_index < len || len == 0, "index too large"); |
245 |
|
246 |
- const size_t stride = MIN2(len - beg_index, ObjArrayMarkingStride); |
247 |
+ const size_t stride = MIN2(len - beg_index, (size_t)ObjArrayMarkingStride); |
248 |
const size_t end_index = beg_index + stride; |
249 |
T* const base = (T*)a->base(); |
250 |
T* const beg = base + beg_index; |
251 |
diff -up jdk8/hotspot/src/share/vm/runtime/arguments.cpp.size_t jdk8/hotspot/src/share/vm/runtime/arguments.cpp |
252 |
--- jdk8/hotspot/src/share/vm/runtime/arguments.cpp.size_t 2014-03-11 16:02:13.000000000 -0400 |
253 |
+++ jdk8/hotspot/src/share/vm/runtime/arguments.cpp 2014-05-16 13:26:32.310132543 -0400 |
254 |
@@ -1240,7 +1240,7 @@ void Arguments::set_cms_and_parnew_gc_fl |
255 |
// NewSize was set on the command line and it is larger than |
256 |
// preferred_max_new_size. |
257 |
if (!FLAG_IS_DEFAULT(NewSize)) { // NewSize explicitly set at command-line |
258 |
- FLAG_SET_ERGO(uintx, MaxNewSize, MAX2(NewSize, preferred_max_new_size)); |
259 |
+ FLAG_SET_ERGO(uintx, MaxNewSize, MAX2((size_t)NewSize, preferred_max_new_size)); |
260 |
} else { |
261 |
FLAG_SET_ERGO(uintx, MaxNewSize, preferred_max_new_size); |
262 |
} |
263 |
@@ -1265,8 +1265,8 @@ void Arguments::set_cms_and_parnew_gc_fl |
264 |
// Unless explicitly requested otherwise, make young gen |
265 |
// at least min_new, and at most preferred_max_new_size. |
266 |
if (FLAG_IS_DEFAULT(NewSize)) { |
267 |
- FLAG_SET_ERGO(uintx, NewSize, MAX2(NewSize, min_new)); |
268 |
- FLAG_SET_ERGO(uintx, NewSize, MIN2(preferred_max_new_size, NewSize)); |
269 |
+ FLAG_SET_ERGO(uintx, NewSize, MAX2((size_t)NewSize, min_new)); |
270 |
+ FLAG_SET_ERGO(uintx, NewSize, MIN2(preferred_max_new_size, (size_t)NewSize)); |
271 |
if (PrintGCDetails && Verbose) { |
272 |
// Too early to use gclog_or_tty |
273 |
tty->print_cr("CMS ergo set NewSize: " SIZE_FORMAT, NewSize); |
274 |
@@ -1276,7 +1276,7 @@ void Arguments::set_cms_and_parnew_gc_fl |
275 |
// so it's NewRatio x of NewSize. |
276 |
if (FLAG_IS_DEFAULT(OldSize)) { |
277 |
if (max_heap > NewSize) { |
278 |
- FLAG_SET_ERGO(uintx, OldSize, MIN2(NewRatio*NewSize, max_heap - NewSize)); |
279 |
+ FLAG_SET_ERGO(uintx, OldSize, MIN2((size_t)(NewRatio*NewSize), max_heap - NewSize)); |
280 |
if (PrintGCDetails && Verbose) { |
281 |
// Too early to use gclog_or_tty |
282 |
tty->print_cr("CMS ergo set OldSize: " SIZE_FORMAT, OldSize); |
283 |
@@ -1397,7 +1397,7 @@ bool verify_object_alignment() { |
284 |
return true; |
285 |
} |
286 |
|
287 |
-uintx Arguments::max_heap_for_compressed_oops() { |
288 |
+size_t Arguments::max_heap_for_compressed_oops() { |
289 |
// Avoid sign flip. |
290 |
assert(OopEncodingHeapMax > (uint64_t)os::vm_page_size(), "Unusual page size"); |
291 |
// We need to fit both the NULL page and the heap into the memory budget, while |