-
Notifications
You must be signed in to change notification settings - Fork 426
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Input_shape (None, None, 3) #98
Comments
With
The https://www.tensorflow.org/api_docs/python/tf/image/resize_images#args |
If I use
which cause error:
|
In the end, I have no choice but reload model every time before taking a new image as input:
|
Wait for a minute. Maybe we don't need this feature at all. Because normally, you'll deal same resolution images at one specific task. For example, process a whole length of movie or video. (Or you can still use a fixed resolution image as input, then scale the resulting mask by yourself. |
That's a workaround indeed. Resizing might affect segmentation performance though, as object dimensions change. Moreover, it's computationally expensive when doing on the fly or you need a second copy of your images on hard disk if you wanna do it up front. That's why flexible model inputs are desirable I think. |
Referring to closed issue #65:
I still can't load the model with input_shape=(None, None, 3) using the latest (TF2-based) version of this repo:
Deeplabv3(weights='pascal_voc', input_shape=(None, None, 3), classes=1, backbone='xception', OS=16)
gives me
caused by lines 417-419 in model.py. I can load the model when replacing x.shape with tf.shape(x) in line 418, but then I'm getting another shape error during training (model.fit). Am I doing something wrong? Do others get the same error message? I tried to solve the issue on my own, but it seems to be beyond my skills.
Additional info: using input_shape of (248,248,3) causes
in Line 425 in model.py. The TF1.13-based version of this repo accepts 248*248 input images.
The text was updated successfully, but these errors were encountered: